Skip to main content
PeerJ logoLink to PeerJ
. 2020 Jul 20;8:e9539. doi: 10.7717/peerj.9539

The importance of open science for biological assessment of aquatic environments

Marcus W Beck 1,2,, Casey O’Hara 3, Julia S Stewart Lowndes 4, Raphael D Mazor 1, Susanna Theroux 1, David J Gillett 1, Belize Lane 5, Gregory Gearheart 6
Editor: Robert Toonen
PMCID: PMC7377246  PMID: 32742805

Abstract

Open science principles that seek to improve science can effectively bridge the gap between researchers and environmental managers. However, widespread adoption has yet to gain traction for the development and application of bioassessment products. At the core of this philosophy is the concept that research should be reproducible and transparent, in addition to having long-term value through effective data preservation and sharing. In this article, we review core open science concepts that have recently been adopted in the ecological sciences and emphasize how adoption can benefit the field of bioassessment for both prescriptive condition assessments and proactive applications that inform environmental management. An example from the state of California demonstrates effective adoption of open science principles through data stewardship, reproducible research, and engagement of stakeholders with multimedia applications. We also discuss technical, sociocultural, and institutional challenges for adopting open science, including practical approaches for overcoming these hurdles in bioassessment applications.

Keywords: Applied science, Bioassessment, Open data, Open science, Reproducible, R

Introduction

Bioassessment is an essential element of aquatic monitoring programs that helps guide decisions for managing the ecological integrity of environmental resources. Legal mandates to assess biological condition have stimulated the development of bioassessment programs and tools in the United States (Clean Water Act, CWA), Canada (Canada Waters Act), Europe (Water Framework Directive), China (Environmental Quality Standards for Surface Water), South Africa (National Water Act), and elsewhere (Borja et al., 2008). Decades of research have supported the development of assessment indices for multiple assemblages with regional applications in streams, rivers, lakes, and marine environments (Karr et al., 1986; Kerans & Karr, 1994; Fore & Grafe, 2002; Beck & Hatch, 2009; Borja, Ranasinghe & Weisberg, 2009; Borja et al., 2016). Substantial technical advances have been made in measuring biological responses to environmental change (Hawkins et al., 2000a; Hawkins et al., 2000b), how these responses can be distinguished from natural environmental variation (Stoddard et al., 2006; Hawkins, Olson & Hill, 2010), and interpreting the impacts of these changes (Davies & Jackson, 2006).

Integrating bioassessment products (e.g., scoring indices, causal assessment protocols) into management or regulatory frameworks can be challenging, despite the technological advances (Kuehne et al., 2017). How a bioassessment product is used in practice to inform decisions and prioritize management actions can differ from why it may have been originally developed. Numerous assessment products have been developed for specific regional applications (Birk et al., 2012) and concerns about redundancy, comparability, duplicated effort, and lack of coordinated monitoring have recently been highlighted (Cao & Hawkins, 2011; Poikane et al., 2014; Kelly et al., 2016; Nichols et al., 2016). Kuehne, Strecker & Olden (2019) recently highlighted a lack of institutional connectivity among actors with expertise in freshwater assessment as a hallmark of the status quo in which applied science is conducted. Moreover, existing indices may not be easily calculated by others beyond initial research applications (Hering et al., 2010; Nichols et al., 2016) or may be incorrectly applied based on differences between goals for developing an index and the needs of management programs (Dale & Beyeler, 2001; Stein et al., 2009). The abundance of available products can be a point of frustration for managers given a lack of guidance for choosing among alternatives, particularly as to how different assessment products relate to specific management, monitoring, or policy objectives (Dale & Beyeler, 2001; Stein et al., 2009).

To address these challenges, a new mode of operation is needed where method development is open and transparent, developed products are discoverable and reproducible, and most importantly, implementation in the management community is intuitive and purposeful. Open science principles that improve all aspects of the scientific method can help meet these needs and there is a unique opportunity in bioassessment to leverage openness to support public resources. Open science and its ideals originated partly due to failures of reproducibility and biases in the primary literature that were revealed as systematic concerns in research fields with immediate implications for human health (Makel, Plucker & Hegarty, 2012; Franco, Malhotra & Simonovits, 2014). These ideas and the failures that they address have slowly permeated the ecological and environmental sciences (Hampton et al., 2015; Hampton et al., 2016; Lowndes et al., 2017). Open science has also influenced how research workflows are conceptualized in other disciplines (e.g., archaeology, Marwick et al., 2016, behavioral ecology, Ihle et al., 2017, hydrology, Slater et al., 2019, vegetation sciences, Collins, 2016) and has enabled a shift towards publishing structures that are more fair and transparent through open access (Van Oudenhoven, Schröter & De Groot, 2016; Essl et al., 2020). Limited examples have suggested that open access databases can be leveraged to develop bioassessment products that increase transparency among stakeholders (Borja et al., 2019). Adopting an open science paradigm in bioassessment is particularly relevant compared with other fields given the explicit need to develop products that are accessible to the management community. Legal and ethical precedents in bioassessment may also necessitate open data sharing given that environmental monitoring programs are often established to protect and maintain publicly-owned natural resources.

Survey Methodology and Objectives

This review draws on previous literature to describe approaches for open science that can empower the research and management community to embrace a new mode of thinking for bioassessment applications. These approaches are expected to benefit the bioassessment research community by providing new tools that augment existing workflows for developing assessment products and improving their ability to address environmental issues by bridging the gap between the scientific, management, and regulatory communities. The intended audience for this review is primarily the research team that develops bioassessment products, but we also write for the funders and users (e.g., regulators and managers) of these products to emphasize the value of investing in open science for the protection of public resources.

This traditional review covers literature published in recent years advocating for open science in different fields of study. Because no similar efforts have yet been made to apply these principles to bioassessment, we draw on examples from the previous literature that demonstrate successful applications in other fields to motivate researchers and practitioners to embrace these new ideas in bioassessment. Comprehensive and unbiased coverage of the previous literature was accomplished by querying online search engines, primarily Google Scholar, with search terms as they relate to open science (e.g., “reproducibility”, “data science”, “open source”) and with Boolean operators to find applications to bioassessment (e.g., “reproducibility AND bioassessment”). Studies were included if they provided general overview of open science concepts that were relevant to bioassessment or if they directly described open science applications to bioassessment, although the latter were scarce. Emphasis was given to the breadth of research that has supported the development of open source software applications that can aid bioassessment, both as general tools and more specific programs tailored for indicator development. We excluded studies that described applications with citizen science components. Although citizen science can be a valuable tool for researchers and managers, methods for effective implementation are beyond the scope of this review.

Our objectives are to (1) provide a general overview of principles of open science and (2) empower the research community by providing examples of how these principles can be applied to bioassessment. For the second objective, we also provide a case study of stream bioassessment in the urban landscape of southern California to demonstrate a successful proof of concept. Herein, open science ‘tools’ describe best practices and specific applications that use an open philosophy to support applied science. We structure the review by first introducing open science principles, then describing how these principles could be applied to bioassessment (i.e., developing goals, curating data, and applying open-source software) including a case study example, and lastly providing a discussion of limitations and opportunities to better contextualize real world applications of open science.

Principles of open science

Conventional modes of creating scientific products and more contemporary approaches that align with open science principles share the same goals. Both are motivated by principles of the scientific method that make the process of discovery transparent and repeatable. Where the conventional and open science approaches diverge is the extent to which technological advances facilitate the entire research process. Distinction between the two approaches can be conceptualized as the “research paper as the only and final product” for the conventional approach, whereas the open science approach is inherently linked to advances in communication and analysis that have been facilitated by the Internet and computer sciences (Table 1). As a result, the open science approach can enhance all aspects of the scientific process from initial conception of a research idea to the delivery and longevity of a research product (Fig. 1). The process is iterative where products are improved by the individual and/or others, facilitated by open science tools that enhance access and reproducibility of data.

Table 1. Core definitions and principles of open science.

Content adapted from Open Knowledge International, http://opendefinition.org/, Creative Commons, https://creativecommons.org/about/program-areas/open-science/, D. Gezelter, http://openscience.org/, and Powers & Hampton (2019).

Concepts and principles Description
Open Anyone can freely access, use, modify, and share for any purpose
Open Science Practicing science in such a way that others can collaborate and contribute, where research data, lab notes and other research processes are freely available, under terms that enable reuse, redistribution and reproduction of the research and its underlying data and methods
Reproducible Producing equivalent outcomes from the same data set, or in the case of computational reproducibility, producing equivalent outcomes from the same data set using the same code and software as the original study
Principle 1 Transparency in experimental methods, observations, and collection of data
Principle 2 Public availability and reusability of scientific data
Principle 3 Public accessibility and transparency of scientific communication
Principle 4 The use of web-based tools to facilitate scientific collaboration and reproducibility

Figure 1. A simplified workflow of the open science paradigm (adapted from Hampton et al., 2015).

Figure 1

All aspects of the research process, from the conception of an idea to publishing a product, can be enhanced using open science tools. The workflow is iterative where products are continually improved through collaborations facilitated through discovery and reproducibility of open data.

The paradigm of the research paper as a final scientific product can inhibit the uptake of research methods and findings by environmental managers. The research paper is conventionally viewed as a communication tool for scientists to report and share results among peers. Researchers access periodicals to stay informed of scientific advances and use the information to replicate and improve on methods for follow-up analysis. Although the primary literature continues to serve this critical role, this workflow is problematic when scientific products are needed to serve interests outside of the research community. For example, the paper as an endpoint for environmental managers fails to deliver products that are easily accessible from the practitioner’s perspective, both in application and interpretation. A research paper is less likely to effect environmental change because it does not provide a mechanism to transfer actionable information to those that require scientific guidance for decision-making, such as sharing analysis code or results that describe output from assessment products. Numerous studies have documented implementation failures as a result of siloing among research communities where the flow of information does not extend beyond institutional walls (e.g., Mitchell, 2005; Liu et al., 2008). Information loss over time is another concern associated with the paradigm of research paper as final product (Michener et al., 1997), particularly as intimate knowledge of study details is lost as new projects are initiated or individuals leave institutions.

Open data as a component of the open science process

Open data is a fundamental component of the broader open science process described in Fig. 1. Under this mode of thinking, the research team becomes stewards of its data. For bioassessment data, government institutions may be the primary stewards of information that supports product development within a broader research team. Stewardship allows the data to be treated as a dynamic product with a traceable and replicable provenance (i.e., origin), rather than proprietary and serving only the internal needs of an immediate research goal. Metadata that describe the structure and history of a dataset ensure the data have an identity. Metadata also encourage adoption of core data structures that allow integration across different sources, which is critical for collaboration across institutional boundaries (Horsburgh et al., 2016; Hsu et al., 2017). Other open science practices, such as integration of data with dynamic reporting tools or submitting data to a federated repository (i.e., a decentralized database system for coordination and sharing), can facilitate communication for researchers and those for which the research was developed (Bond-Lamberty, Smith & Bailey, 2016). Prominent examples that can benefit next-generation bioassessment methods, such as molecular-based techniques for species identification, include the BarCode of Life Data Systems (BOLD) and GenBank repositories.

Open data can benefit research by contributing to an increase in novel products created through collaboration. Collaborative publications have increased in the environmental sciences as research teams leverage open data to create synthesis products that allow novel insights from comparisons across multiple datasets. Quantitative meta-analyses and systematic reviews are increasingly used to extract information from the primary literature (Carpenter et al., 2009; Lortie, 2014). In addition, open data products can increase efficiency of the individual researcher and a collective research team by encouraging collaborators to adopt an open science workflow. Many tools developed within the software and computer science community to facilitate open process and the creation of open data are now easily accessible to environmental scientists (Yenni et al., 2019). Version control software (e.g., Git, GitHub), open source programming languages (e.g., R, Python), and integrated development environments (IDEs, e.g., RStudio, Spyder) can all be leveraged to dynamically create and share open data products that can build institutional memory. These tools promote deliberate and shared workflows among researchers that can lead to better science in less time (Lowndes et al., 2017) and have proven useful in recent applications in the hydrologic sciences (Idaszak et al., 2017; Slater et al., 2019).

Open access to data can also benefit management and regulatory communities. Openness can improve the value of data from monitoring programs by facilitating data discovery and synthesis, often through the adoption of a common metadata structure and integration of data within federated data networks (e.g., DataONE, iRODS). Research institutions can also use open data maintained by management or regulatory communities to develop products that directly support the mission of the latter, e.g., assessment methods developed from long-term monitoring datasets that identify priority areas to focus management actions or fulfill regulatory obligations. Open data can also improve public trust in scientific findings by exposing the underlying information used to develop a research product (Grand et al., 2012). Similar concepts are used in “blockchain” technologies that allow public financial transactions in an open, distributed format, as for trading in cryptocurrencies (Pilkington, 2016). Increased trust could facilitate eventual adoption of proposed rules or regulations that are based on research products created from open data. More efficient and effective implementation of potential regulations may also be possible if supporting data are openly available.

Applying open science principles to bioassessment

Here we provide a detailed description of open science processes that the bioassessment community could leverage to create reproducible, transparent, and discoverable research products for environmental managers. The below examples require understanding the distinction between the general open science process in Fig. 1, open data as an individual component of the open science process, and the technology-based tools that can be used to achieve these ends. Both the tools and open data are critical components that facilitate the broader process to achieve the principles outlined in Table 1. “Openness” of process, tools, and data exists on a continuum, and incremental improvements can transform an individual’s and research group’s practice over time. We encourage awareness that an open process adopts the open science tools that are appropriate for a research question and the creation of open data can be a fundamental component of the process. Acceptance by the research team and collaborators of the concepts described in Table 1 is critical to achieving openness.

The overall process is shown in Fig. 2 as an expansion of general concepts in Fig. 1. This iterative flow of information is facilitated by (1) openly sharing planning documents, (2) using established metadata standards to document synthesized data products, (3) hosting data products on open repositories, (4) creating reproducible summary documents that integrate the data and research products, and (5) incorporating the developed product into interactive applications that deliver the results to the managers and stakeholders. The technical phase of defining research goals, collecting and synthesizing data, and developing the bioassessment product are primary tasks of the research team. However, the open science process is distinguished by the flow of information to and from the research phase that can benefit the specific project and the science of bioassessment as a whole.

Figure 2. An idealized open approach for bioassessment applications.

Figure 2

The green box represents the technical steps of the individual research team for developing the product, the manager and stakeholder box are those that require or motivate the creation of bioassessment products, the gray boxes indicate sources of external information (data and guidance documents) as input into the technical process, and the open text indicates open components of the planning, application, or implementation phase of a bioassessment product. The figure was adapted from Hampton et al. (2015). NGO: non-government organization, RMP: regional monitoring program.

Developing bioassessment goals

In an open science process, the goals identified by the research team for developing a bioassessment product should occur through direct, two-way interaction with the management or regulatory institution that requires the product. Although such an approach has historically been used to develop bioassessment products, the interaction in an open science workflow differs in how information is exchanged. This exchange can be accomplished through direct communication and sharing of planning documents to ensure all decisions are transparent, i.e., open planning. In person meetings are ideal, but planning documents are dynamic and will require remote sharing and revision as ideas progress. Online tools such as Google documents, Slack discussion channels, and open lab notebooks can be instrumental for collaboration. More informal approaches, such as blogging and sharing ideas on social media, can expose new concepts to the broader community for guidance (Woelfle, Olliaro & Todd, 2011; Darling et al., 2013). Overall, the research team should use these tools to identify stakeholder needs while also considering the balance between the research goals and limitations of the data to meet these goals. This approach will ensure that the needs of the management and stakeholder communities will be consistent with the services provided by the research product.

An important practice that is not often used in bioassessment for project planning is study pre-registration. This is a relatively new addition to the philosophy of open science that allows a research team to define their study procedures, expected outcomes, and statistical analysis plans in advance of the actual study (Munafó et al., 2017). Although the standard scientific method may seem to support such proactive practices, pre-registration is an explicit declaration to make the intentions of a study design clear to avoid publication bias where only positive outcomes are reported and to prevent an interpretive result where the researcher retrospectively defines study objectives after initial results are obtained if they do not agree with expectations. This latter issue is a serious concern where scientists use postdiction with significant hindsight bias in place of prediction and conventional hypothesis testing to inform scientific discovery (Nosek et al., 2018). Registered reports can also be used as a publishing format where an initial study design is peer-reviewed and the article is provisionally accepted by a journal if the results are created with methods that do not deviate from the accepted design. This promotes greater coverage in the primary literature of null conclusions that otherwise may not have been published, reducing bias for publishing positive results.

Pre-registration has been used extensively in clinical research (Dickersin & Rennie, 2003), where outcomes often have immediate implications for human health and well-being. In contrast, bioassessment studies often focus on developing applied products, where conventional hypothesis testing is less a concern. Studies are typically methods-focused where a research product is developed to address a management or regulatory need, rather than a specific research question with a testable hypothesis. However, pre-registration could be an important tool for the environmental sciences where an explicit declaration of study intent as being applied or methods-focused could prevent postdiction or an otherwise misuse of study results after a project is completed. Existing venues that support pre-registration of studies across multiple disciplines could be used in bioassessment study planning (e.g., Open Science Framework, AsPredicted).

Curating bioassessment data

After project goals are established, the research team identifies requirements and sources of data that need to be synthesized to meet the research needs. Bioassessment data, or more generally, biological data obtained from field sampling have a unique set of challenges that require added vigilance in data stewardship (Cao & Hawkins, 2011). Species identification requires a tradeoff between taxonomic specificity and cost (Lenat & Resh, 2001; Chessman, Williams & Besley, 2007). Species names also change regularly requiring updates to standard taxonomic effort (STE) tables that are critical for many biological indices, although some standardized databases have facilitated broad-scale comparisons (e.g., World Register of Marine Species, Costello et al., 2013). Unidentified or ambiguous taxa must also be explicitly treated in analysis workflows (Cuffney, Bilger & Haigler, 2007), e.g., are they treated as missing values or are they substituted with coarser taxonomic designations? In contrast, molecular techniques based on DNA barcoding eliminate the need for direct species collection and morphological identification (Deiner et al., 2017; Hering et al., 2018). These next-generation approaches have capitalized on advances in database development that allow open access by diverse researchers across disciplines and are well-suited for the development of additional open science tools. Despite these advances, molecular-based approaches have also suffered from challenges related to standardization of workflows and coverage of reference databases (White et al., 2014; Elbrecht et al., 2017).

Open science tools can facilitate the curation of bioassessment data by addressing the above challenges. For example, a multimetric index may require taxonomic data collected at multiple sites by different institutions, whereas the output data may include summary scores, individual metrics, and any additional supporting information to assess the quality of the output. In an open science workflow, these data products can be documented using a standardized metadata language (e.g., Ecological Metadata Language Standard, or EML) which describes the who, what, and why to ensure the data have an identity. Adoption of a metadata standard also ensures that a machine-readable file is produced to allow integration into a data repository. This will allow a synthesized data product to be discoverable beyond the specific research application and will provide metadata to help others understand the context of the data (e.g., Idaszak et al., 2017). Finally, the dataset can be assigned a unique digital object identifier (DOI, e.g., through Zenodo) that provides a permanent address and is also citable to allow researchers to track usage of a bioassessment data product.

In an open paradigm, the data itself is a product to achieve the research goals and also becomes available to the research and management community as a fully documented source of information that has value beyond the specific project. The openness of the synthesized data product is one of the primary means of facilitating the application of a bioassessment product. The synthesized data product can be used by the research team to create interactive applications for stakeholders to share and explore the data and is also fully integrated into summary reports using software for generating dynamic documents (e.g., using knitr, Xie, 2015, RMarkdown, Allaire et al., 2018, Jupyter notebooks, Kluyver et al., 2016). Continuous integration services can automate quality control and regularly update data products as new information is collected (Yenni et al., 2019). The data product also becomes available on an open data repository that is discoverable by other researchers and can contribute to alternative scientific advances beyond the immediate goals (e.g., Hydroshare for the hydrologic sciences, Idaszak et al., 2017).

Using R for bioassessment application

The R statistical programming language (RDCT, 2020) is one of the most commonly used analysis platforms in the environmental sciences (Lai et al., 2019; Slater et al., 2019) and many existing R packages have value for the bioassessment community (Table 2). For managing the day to day tasks of working with multiple datasets, the tidyverse suite of packages provides the necessary tools to import, wrangle, explore, and plot almost any data type (Wickham, 2017). The tidyverse also includes the powerful ggplot2 package that is based on a syntactical grammar of graphics for plotting (Wilkinson, 2005; Wickham, 2009). This package provides a set of independent plotting instructions that can be built piecewise and is a departure from other graphics packages that represent a collection of special cases that limit the freedom of the analyst. In bioassessment, ggplot2 can be used both in an exploratory role during the development phase and also to create publication quality graphics.

Table 2. R packages that can be used in the development and application of bioassessment products.

Task Package Description
General tidyverse (Wickham, 2017) A suite of packages to import, wrangle, explore, and plot data. Includes the popular ggplot2 and dplyr packages.
Mapping, geospatial sf (Pebesma, 2018) A simple features architecture for working with vectorized spatial data, including common geospatial analysis functions
raster (Hijmans, 2019) Reading, writing, manipulating, analyzing, and modeling gridded spatial data
leaflet (Cheng, Karambelkar & Xie, 2018) Integration of R with the popular JavaScript leaflet library for interactive maps
mapview (Appelhans et al., 2018) Creates interactive maps to quickly examine and visually investigate spatial data, built off leaflet and integrated with sf
Statistical modeling randomForest (Liaw & Wiener, 2002) Create classification and regression trees for predictive modeling
nlme (Pinheiro et al., 2018) Non-linear, mixed effects modeling
mgcv (Wood, 2017) Generalized additive modeling
Community analysis TITAN2 (Baker, King & Kahle, 2015) Ecological community threshold analysis using indicator species scores
indicspecies (De Caceres & Legendre, 2009) Indicator species analysis
vegan (Oksanen et al., 2018) Multivariate analysis for community ecology
Science communication shiny (Chang et al., 2018) Reactive programming tools to create interactive and customizable web applications
rmarkdown (Allaire et al., 2018) Tools for working with markdown markup languages in .Rmd files
knitr (Xie, 2015) Automated tools for markdown files that process integrated R code chunks

Bioassessment data are inherently spatial and recent package development has greatly improved the ability to analyze and map geospatial data in R. The raster package can used to read/write, manipulate, analyze, and model grid-based spatial data (Hijmans, 2019), which are often common supporting layers for bioassessment (e.g., elevation or climate data). For vector data (i.e., points, lines, and polygons), the sf package (“simple features”, Pebesma, 2018) uses principles of data storage that parallel those from the tidyverse by representing spatial objects in a tidy and tabular format. This facilitates analysis by presenting complex spatial structures in a readable format that can be integrated in workflows with existing packages, including other mapping packages (e.g., leaflet, Cheng, Karambelkar & Xie, 2018, or mapview, (Appelhans et al., 2018)). This allows the research team to use a workflow that is focused in a single environment, rather than using separate software for statistical and geospatial analysis.

R is fundamentally a statistical language and several existing R packages can be used to evaluate and support bioassessment data. Random forest models have been used to develop predictive bioassessment indices that compare observed taxa to modeled expectations (i.e., O/E indices). The randomForest package (Liaw & Wiener, 2002) uses an ensemble learning approach that is robust to complex, non-linear relationships and interactions between variables. These models are particularly useful with large, regional datasets that describe natural and anthropogenic gradients in condition (Laan & Hawkins, 2014; Mazor et al., 2016). The nlme package can be used to create non-linear mixed effect models that are more flexible than standard regression approaches (Pinheiro et al., 2018). The nlme package can develop models for nested sampling designs, such as repeat visits to sample sites or otherwise confounding variables that contribute information but are not unique observations (Mazor et al., 2014). The mgcv package provides similar functionality as nlme, but uses an additive modeling approach where individual effects can be evaluated as the sum of smoothed terms (Wood, 2017). The mgcv package is often applied to model biological response to stressor gradients (Yuan, 2004; Taylor et al., 2014).

Other R packages have been developed specifically for bioassessment. For example, the TITAN2 package can be used to develop quantitative evidence of taxon-specific changes in abundance and occurrence across environmental gradients (Baker, King & Kahle, 2015). Results from this package can support exploratory analysis for developing bioassessment products, such as identifying indicator species or establishing numeric criteria (Taylor et al., 2018). The results can be also be used post hoc to evaluate potential response of a biological index with changing environmental conditions, such as proposed management actions for rehabilitation (King et al., 2011). Alternatively, the indicspecies package provides similar functionality but is based only on species occurrence or abundance matrices across sites (De Caceres & Legendre, 2009). This package can be used to identify species at sites if continuous environmental data are unavailable, such as those that are representative of reference conditions (Bried et al., 2014). Finally, the vegan package has been a staple among community ecologists for multivariate analyses in R (Oksanen et al., 2018).

Although the R network includes over 15,000 user contributed packages, only a handful of these packages are specific to bioassessment. Community practices have allowed R to reach new audiences where new packages build on the work of others and are transportable between users and operating systems. Formalized communities, such as rOpenSci, encourage standardization and review of contributed packages within the ecological sciences to make scientific data retrieval reproducible. Several tools have also been developed and published in the last five years that greatly simplify the process of creating new packages in R (Wickham, 2015; Wickham, Hester & Chang, 2018). The advantages of creating and sharing R packages that are specific to bioassessment applications are important for several reasons. First, an R package compartmentalizes technical instructions developed during the research phase that can be executed by anyone with access to the software. R packages also require explicit documentation of the functions and data requirements. As such, package users will not only have access to underlying code but also understand the why and what for different package functions.

Finally, R can be used to create interactive applications that deliver bioassessment products to stakeholders and managers in entirely novel contexts. In particular, the shiny package provides programming tools built around concepts of reactivity, where data inputs and outputs can be modified in real time (Chang et al., 2018). A shiny application is an interactive user interface that is developed with R code, but is a standalone product that can be used without any programming experience. These applications are deployed online and can extend the reach of bioassessment products to those that require the information for decision-making but otherwise do not have the time or resources to learn R. Applications built in shiny can also be easily linked to other R packages. For example, a shiny website could be created to allow users to upload raw data and estimate and report bioassessment scores using an R package developed externally. Moreover, shiny applications are completely customizable and can be tailored by the developer to the specific needs of any user. This distinction separates shiny from other web-based analysis platforms.

Open science in practice: the SCAPE project

Although bioassessment products have been sufficiently developed in California (USA), there are no narrative or numeric criteria in place to support designated aquatic life uses in wadeable streams, nor are bioassessment data actively used to support conservation or watershed management. Indices using benthic macroinvertebrates and algae have been developed that provide consistent indications of biological condition across the diverse geography and climates in the state (Fetscher et al., 2013; Mazor et al., 2016; Ode et al., 2016). A physical habitat index has also been developed that provides complementary information supporting bioassessment data (Rehn, Mazor & Ode, 2018). Combined, these indices represent significant achievements in overcoming technical challenges for developing accurate and interpretable bioassessment products. However, these products are not used at a statewide scale to inform decisions and past efforts for stream management have only used a fraction of available products. A synthesis of condition assessments is needed to effectively implement bioassessment products in California and data must be presented in a context that is relevant to the needs of decision makers.

Recent regulatory initiatives in California have established a foundation for openness that could greatly improve the application of bioassessment products to support decision-making. In particular, these initiatives have set a precedent for openly sharing data collected with public funds. The Open and Transparent Water Data Act passed by the state legislature in 2016 requires water quality institutions to “create, operate, and maintain a statewide integrated water data platform that, among other things, would integrate existing water and ecological data information from multiple databases and provide data on completed water transfers and exchanges” (AB 1755, Dodd, 2015--2016). This legislation also calls for state agencies to “develop protocols for data sharing, documentation, quality control, public access, and promotion of open-source platforms and decision support tools related to water data”. These aspirations were further supported by a resolution on July 10, 2018 that formally committed the State Water Resources Control Board to “provide broader access to data used to make local, regional, and statewide water management and regulatory decisions in California”. These recent initiatives in California have similarly been observed at the national level. For example, the Data Coalition is an advocacy group that operates on behalf of the private and public sector for the publication of government data in a standardized and open format. The Internet of Water also operates at the national-level by focusing on strengthening connections between data producers and users through centralized data hubs and data standards.

Open science tools have recently been used in California to address bioassessment implementation challenges in developed landscapes. The Stream Classification and Priority Explorer, or SCAPE (Beck, 2018a; Beck et al., 2019), was developed using an open science framework to help identify reasonable management goals for wadeable streams using existing bioassessment and watershed data. The SCAPE tool represents both a modeling approach to help prioritize management goals (Fig. 3) and a set of open science products for direct application to environmental managers. The modeling component addresses a practical problem of achieving reference conditions in developed landscapes, where channel modification is common. Using the National Hydrography Dataset (NHD-Plus; McKay et al., 2012) and watershed predictors (StreamCat; Hill et al., 2016), the model classifies stream segments as biologically “constrained” or “unconstrained” by landscape alteration. This classification system can be used to set management priorities based on the constraint class. For example, a monitoring site with an observed biological index score that is above a predicted range could be assigned a higher management priority relative to a site that is scoring within the range that is expected based on landscape development.

Figure 3. Schematic demonstrating how the Stream Classification and Community Explorer (SCAPE) can be used to identify potential management actions for stream sites.

Figure 3

Stream segment classifications are defined as biologically constrained or unconstrained based on landscape characteristics (A) and sites with bioassessment scores are evaluated relative to the classifications. Sites can be under-scoring, as expected, or over-scoring relative to the segment classification and expected range of scores (B). Unconstrained sites are those where present landscape conditions do not limit biological potential and constrained sites are those where landscape conditions limit biological potential. Management actions and priorities can be defined based on site scores relative to segment classifications. TMDL: Total Maximum Daily Load. Photo credit: Raphael Mazor.

Open science tools were critically important for translating and delivering SCAPE products to decision-makers. Local stakeholder engagement to identify research goals guided the technical development process of SCAPE. All analyses, including model development and validation, were conducted using R. A version control system (Git) and online hosting (GitHub) also allowed full transparency of decisions that were made to create the SCAPE model. A permanent DOI was assigned through Zenodo to track downloads and portability of source code (Beck, 2018a). Importantly, an online, interactive web page (https://sccwrp.shinyapps.io/SCAPE) greatly increased the impact and relevance of SCAPE by improving stakeholder understanding through direct interaction with key decision points that influenced model output. A manuscript describing the technical components of the model was created using knitr and RMarkdown (Xie, 2015; Allaire et al., 2018). This increased efficiency of the writing process also minimized the potential of introducing errors into tables or figures by eliminating the need to copy results between different writing platforms. Finally, a geospatial data file from the model was also made public on a federated data repository, which included metadata and plain language documentation to track provenance of the original information (Beck, 2018b).

Limitations and opportunities

Although the case for open science in bioassessment is appealing, the widespread adoption of these principles in practice is inhibited by inertia of existing practices, disciplinary culture, and institutional barriers. Conventional and closed workflows used by many scientists are adopted and entrenched because of ease of use, precedence, and familiarity, yet they can be inefficient, inflexible, and difficult to communicate or replicate. Open science tools can improve analysis, documentation, and implementation through greater flexibility, but they expose research teams to entirely new concepts and skillsets in which they may never have been trained (e.g., Idaszak et al., 2017). Not only are the required skillsets demanding, but the open science toolbox continues to expand as new methods are developed and old methods become obsolete. This requires a research team to stay abreast of new technologies as they are developed and weigh the tradeoffs of adopting different workflows for different research tasks.

Advocates for open science are well aware of the technical challenges faced by individuals and research teams that have never been exposed to the core concepts. Most importantly, education and training (e.g., through The Carpentries) remain key components for developing skillsets among researchers where the focus is both on learning new skills for transferability and realizing their value for improving science as a whole (Hampton et al., 2017). A goal of many training curricula is to instill confidence in new users by developing comfort with new workflows, such as replacing a point-and-click style of analysis with one focused on using a command line through a computer terminal. Other approaches to demonstrate the value of new techniques use a side by side approach of closed vs open workflows to show the increased efficiency and power of the latter. Adoption becomes much more reasonable once users realize the value of investing in learning a new skill.

Advocates of open science also recognize the limitations of teaching in that not all audiences can be reached and not all materials are retained or even used after training. A strategy of empowering trainees to become trainers and teach others at their home institutions (e.g., train-the-trainer workshops and programs) enables open science to reach more individuals, and benefits science more broadly as they develop technical and communication skills, and build local communities. Those that also adopt new workflows through training can also direct their research products to facilitate collaboration with non-adopters rather than the latter synthesizing and analyzing their data in potentially suboptimal ways (Touchon & McCoy, 2016). These “champions” can be a voice of encouragement for others by demonstrating how new tools can be introduced and learned over time through shared experiences (Lowndes et al., 2017). This also encourages the development of a community of practice that shares and learns together to navigate the collection of existing and developing open science tools (Stevens et al., 2018). Champions of open science should also be vocal proponents that spread awareness of the value of open science tools, particularly to those that make decisions on project resources. Department heads or administrative leaders may not be aware of the value of investing in open science, particularly if the consequences of not doing so are externalized in indirect costs that are not budgeted. A change in mindset may be needed where open science is viewed as a core tool that is critical to maintaining relevance of a research program in the future (Hampton et al., 2017).

Many scientists feel they cannot prioritize learning new skills given existing demands on their time, particularly if the benefits of these approaches, such as the value for the research team of sharing their data, are not apparent or immediate. Short-term funding and even political cycles can disincentivize scientists from spending time on anything but contractually obligated deliverables, which as noted above, may not effectively apply science in decision-making. This is an acute concern for early career scientists that have higher demands on establishing reputation and credentials, where investments in open science may be seen as detracting from progress (Allen & Mehler, 2019). As an alternative, a practical solution is to actively encourage the investment in open science within the research team or lab, as opposed to placing the burden on the individual as an isolated researcher (i.e., team science, Cheruvelil & Soranno, 2018). Laboratory or department heads should allow and encourage research staff to invest time in learning new skills and exploring new ideas, even if this does not immediately benefit the latest project. Over time, small investments in developing new skills will have long-term payoffs, particularly if a growing skillset among the research team encourages networking and peer instruction (Lowndes et al., 2017; Allen & Mehler, 2019). Developing an environment where open science tools are highly valued and encouraged may also increase job satisfaction and benefit recruitment and retention if researchers are allowed the space and time to develop skills beyond the current project.

The scientific culture within a discipline or institution may inhibit the adoption of open science methods. A common argument against open science is the protection of data that an individual research team may view as proprietary or sensitive. There are reasonable arguments to treat data as personal property, particularly if exceptional effort was spent to secure funding for a project and if the data were hard-earned or sensitive, e.g., detailed location data on endangered species or medical/socioeconomic data (Zipper et al., 2019). These issues are less of a concern for bioassessment where many datasets are collected by institutions that are publicly funded and data accessibility may be mandated by law. However, an open science process dictates that both interim and completed research products derived from public data should be available to the broader bioassessment community. This raises an additional concern that research teams using transparent workflows could expose themselves to increased criticism by their peers and the public (Lewandowsky & Bishop, 2016; Allen & Mehler, 2019), particularly where the developed products can have important regulatory implications.

Feedback and criticism are fundamental and natural parts of the scientific process. Scientists receive feedback at many stages in the conventional scientific workflow (e.g., internal review, peer-review, presentations at conferences). Potentially new and challenging avenues for feedback are created in an open workflow. A concern is that openness can provide a platform for antagonistic or even hostile views, which could alter or degrade the scientific product (Landman & Glantz, 2009; Lewandowsky & Bishop, 2016). However, opportunities for addressing alternative viewpoints are critical to the open process of creating applied products, even if some voices are politically charged. This is especially true in bioassessment where finished products that could be adopted in regulation are often heavily scrutinized. It is in the interest of applied scientists to hear the concerns of all parties during the development phase. This is not to provide an avenue to erode the integrity or objectives of the science, but to enable full knowledge of the very real barriers to adoption that exist when science is applied in regulation. Openness that invites all voices to participate is a much more agreeable path to consensus than producing the science in isolation of those that it affects (Pohjola & Tuomisto, 2011). Ultimately, these products are developed to improve the environment as a public resource and the ideals promoted by an open science process directly align with these goals.

Institutional barriers can inhibit open science given the scale of change that must occur for adoption. Bureaucratic hurdles can disincentivize initiatives that promote change, particularly if that change originates from researchers not in administrative roles. Regulatory institutions may also prefer some level of opacity for how research products that influence policy are made available during development. The level of transparency advocated by open science could be viewed as opening the floodgates to increased legal scrutiny that can unintentionally hinder forward progress. Despite these reservations, many public institutions now advocate for increased openness because of the benefits that facilitate and engender public trust. Open data initiatives are now fairly common and represent a form of advocacy by public institutions for broader adoption of open science principles. Many national-level data products already exist that embrace openness to invest in the quality and availability of data (e.g., initiatives, US Geological Survey products through NWIS and BioData, US Environmental Protection Agency through STORET/WQX). Internationally, institutions in Europe and Canada that have projects supported by public funds are obligated to publish data and papers as open access (Horizon 2020, Tri-Agency Open Access Policy). Although past efforts and recent changes represent progress, many institutions have yet to strictly define open science and how it is applied internally and externally. As open science continues to build recognition, means of integrating toolsets that promote openness and transparency beyond publicly shared data will have to be adopted by regulatory and management institutions.

Conclusions

The relevance of bioassessment applications can be improved with open science by using reproducible, transparent, and effective tools that bridge the gap between research and management. Many open science tools can improve communication between researchers and managers to expose all aspects of the research process and facilitate implementation to support policy, regulation, or monitoring efforts. Communication ensures that the developed product is created through an exchange of ideas to balance the potentially competing needs of different sectors and institutions. The documentation and archiving of data used to create a bioassessment product also ensures that other researchers can discover and build on past efforts, rather than constantly rebuilding the wheel. Incremental improvements of existing products can reduce the proliferation of site- and taxon-specific methods with limited regional applications by exploring new ways to integrate biological indicators across space and time.

Efforts to formally recognize and integrate open science in bioassessment are needed now more than ever. The transition of bioassessment from taxonomic-based indices to molecular approaches presents novel challenges that will only increase in severity as researchers continue to refine methods for molecular applications (Baird & Hajibabaei, 2012). Although molecular-based indices share similar assessment objectives as conventional indices, the data requirements and taxonomic resolution are substantially more complex. Bioassessment researchers developing molecular methods are and will continue to be inundated with data from high-throughput DNA sequencers. Systematic approaches to document, catalog, and share this information will be required to advance and standardize the science. Molecular approaches are also dependent on existing reference libraries for matching DNA samples for taxonomic identification. The integrity of reference libraries depends greatly on the quality of metadata and documentation for contributed samples. Open science principles should be leveraged in this emerging arena to ensure that new bioassessment methods continue to have relevance for determining the condition of aquatic resources.

Acknowledgments

The authors acknowledge technical support through an advisory committee from the San Gabriel River Regional Monitoring Program and the Sanitation Districts of Los Angeles County. We thank Eric Stein for reviewing an earlier draft of the manuscript. We thank Mike McManus for assistance with the conceptual diagrams. We thank two anonymous reviewers for providing excellent feedback that improved the manuscript.

Funding Statement

The authors received no funding for this work.

Additional Information and Declarations

Competing Interests

The authors declare there are no competing interests. MW Beck is employed by the Tampa Bay Estuary Program, but completed the manuscript as an employee of the Southern California Coastal Water Research Project (SCCWRP). Other non-academic affiliations include RD Mazor, S Theroux, DJ Gillett employed by SCCWRP and G Gearheart employed by the California State Water Resources Control Board.

Author Contributions

Marcus W. Beck and Casey O’Hara conceived and designed the experiments, analyzed the data, prepared figures and/or tables, authored or reviewed drafts of the paper, and approved the final draft.

Julia S. Stewart Lowndes conceived and designed the experiments, prepared figures and/or tables, and approved the final draft.

Raphael Mazor conceived and designed the experiments, analyzed the data, prepared figures and/or tables, and approved the final draft.

Susanna Theroux and David Gillett conceived and designed the experiments, authored or reviewed drafts of the paper, and approved the final draft.

Belize Lane performed the experiments, authored or reviewed drafts of the paper, and approved the final draft.

Gregory Gearheart performed the experiments, analyzed the data, authored or reviewed drafts of the paper, and approved the final draft.

Data Availability

The following information was supplied regarding data availability:

All source materials for reproducing the paper are available in GitHub: https://github.com/fawda123/bioassess_opensci.

References

  • Allaire et al. (2018).Allaire J, Xie Y, McPherson J, Luraschi J, Ushey K, Atkins A, Wickham H, Cheng J, Chang W. Rmarkdown: dynamic documents for R. https://cran.r-project.org/web/packages/rmarkdown/index.html 2018
  • Allen & Mehler (2019).Allen C, Mehler DMA. Open science challenges, benefits and tips in early career and beyond. PLOS Biology. 2019;17:e3000246. doi: 10.1371/journal.pbio.3000246. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • Appelhans et al. (2018).Appelhans T, Detsch F, Reudenbach C, Woellauer S. Mapview: interactive viewing of spatial data in R. https://github.com/r-spatial/mapview GitHub. 2018
  • Baird & Hajibabaei (2012).Baird DJ, Hajibabaei M. Biomonitoring 2.0: a new paradigm in ecosystem assessment made possible by next-generation dna sequencing. Molecular Ecology. 2012;21:2039–2044. doi: 10.1111/j.1365-294X.2012.05519.x. [DOI] [PubMed] [Google Scholar]
  • Baker, King & Kahle (2015).Baker ME, King RS, Kahle D. TITAN2: Threshold indicator taxa analysis. https://CRAN.R-project.org/package=TITAN2 2015
  • Beck (2018a).Beck MW. SCCWRP/SCAPE: v1.0 (Version 1.0). Zenodo. http://doi.org/10.5281/zenodo.1218121 2018a
  • Beck (2018b).Beck MW. Constrained streams for biological integrity in California. https://knb.ecoinformatics.org/view/urn:uuid:75411f50-32ed-42a5-bbfd-26833c7a441f Knowledge Network for Biocomplexity. 2018b
  • Beck & Hatch (2009).Beck MW, Hatch LK. A review of research on the development of lake indices of biotic integrity. Environmental Reviews. 2009;17:21–44. doi: 10.1139/A09-001. [DOI] [Google Scholar]
  • Beck et al. (2019).Beck MW, Mazor RD, Johnson S, Wisenbaker K, Westfall J, Ode PR, Hill R, Loflen C, Sutula M, Stein ED. Prioritizing management goals for stream biological integrity within the developed landscape context. Freshwater Science. 2019;38:883–898. doi: 10.1086/705996. [DOI] [Google Scholar]
  • Birk et al. (2012).Birk S, Bonne W, Borja A, Brucet S, Courrat A, Poikane S, Solimini A, van de Bund W, Zampoukas N, Hering D. Three hundred ways to assess Europe’s surface waters: an almost complete overview of biological methods to implement the Water Framework Directive. Ecological Indicators. 2012;18:31–41. doi: 10.1016/j.ecolind.2011.10.009. [DOI] [Google Scholar]
  • Bond-Lamberty, Smith & Bailey (2016).Bond-Lamberty B, Smith AP, Bailey V. Running an open experiment: transparency and reproducibility in soil and ecosystem science. Environmental Research Letters. 2016;11:1–7. [Google Scholar]
  • Borja et al. (2008).Borja A, Bricker SB, Dauer DM, Demetriades NT, Ferreira JG, Forbes AT, Hutchings P, Jia X, Kenchington R, Marques JC, Zhu C. Overview of integrative tools and methods in assessing ecological integrity in estuarine and coastal systems worldwide. Marine Pollution Bulletin. 2008;56:1519–1537. doi: 10.1016/j.marpolbul.2008.07.005. [DOI] [PubMed] [Google Scholar]
  • Borja et al. (2016).Borja A, Elliot M, Anderson JH, Berg T, Carstensen J, Halpern BS, Heiskanen AS, Korpinen S, Lowndes JSS, Martin G, Rodriguez-Ezpeleta N. Overview of integrative assessment of marine systems: the ecosystem approach in practice. Frontiers in Marine Science. 2016;3:1–20. [Google Scholar]
  • Borja et al. (2019).Borja A, Garmendia JM, Menchaca I, Uriarte A, Sagarmínaga Y. Yes, we can! Large-scale integrative assessment of european regional seas, using open access databases. Frontiers in Marine Science. 2019;6:1–19. [Google Scholar]
  • Borja, Ranasinghe & Weisberg (2009).Borja A, Ranasinghe A, Weisberg SB. Assessing ecological integrity in marine waters, using multiple indices and ecosystem components: challenges for the future. Marine Pollution Bulletin. 2009;59:1–4. doi: 10.1016/j.marpolbul.2008.11.006. [DOI] [PubMed] [Google Scholar]
  • Bried et al. (2014).Bried JT, Jog SK, Dzialowski AR, Davis CA. Potential vegetation criteria for identifying reference-quality wetlands in the south-central United States. Wetlands. 2014;34:1159–1169. doi: 10.1007/s13157-014-0575-5. [DOI] [Google Scholar]
  • Cao & Hawkins (2011).Cao Y, Hawkins C. The comparability of bioassessments: a review of conceptual and methodological issues. Journal of the North American Benthological Society. 2011;30:680–701. doi: 10.1899/10-067.1. [DOI] [Google Scholar]
  • Carpenter et al. (2009).Carpenter SR, Armbrust EV, Arzberger PW, F. S. C III, Elser JJ, Hackett EJ, Ives AR, Kareiva PM, Leibold MA, Lundberg P, Mangel M, Merchant N, Murdoch WW, Palmer MA, Peters DPC, Pickett STA, Smith KK, Wall DH, Zimmerman AS. Accelerate synthesis in ecology and environmental sciences. BioScience. 2009;59:699–701. doi: 10.1525/bio.2009.59.8.11. [DOI] [Google Scholar]
  • Chang et al. (2018).Chang W, Cheng J, Allaire J, Xie Y, McPherson J. Shiny: Web application framework for R. https://CRAN.R-project.org/package=shiny 2018
  • Cheng, Karambelkar & Xie (2018).Cheng J, Karambelkar B, Xie Y. Leaflet: create interactive web maps with the javascript ‘leaflet’ library. https://CRAN.R-project.org/package=leaflet 2018
  • Cheruvelil & Soranno (2018).Cheruvelil KS, Soranno PA. Data-intensive ecological research is catalyzed by open science and team science. BioScience. 2018;68:813–822. doi: 10.1093/biosci/biy097. [DOI] [Google Scholar]
  • Chessman, Williams & Besley (2007).Chessman B, Williams S, Besley C. Bioassessment of streams with macroinvertebrates: effect of sampled habitat and taxonomic resolution. Journal of the North American Benthological Society. 2007;26:546–565. doi: 10.1899/06-074.1. [DOI] [Google Scholar]
  • Collins (2016).Collins SL. Vegetation science in the age of big data. Journal of Vegetation Science. 2016;27:865–867. doi: 10.1111/jvs.12459. [DOI] [Google Scholar]
  • Costello et al. (2013).Costello MJ, Bouchet P, Boxshall G, Fauchald K, Gordon D, Hoeksema BW, Poore GCB, Van Soest RWM, Stöhr S, Walter TC, Vanhoorne B, Decock W, Appeltans W. Global coordination and standardisation in marine biodiversity through the World Register of Marine Species (WoRMS) and related databases. PLOS ONE. 2013;8:e51629. doi: 10.1371/journal.pone.0051629. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • Cuffney, Bilger & Haigler (2007).Cuffney TF, Bilger MD, Haigler AM. Ambiguous taxa: effects on the characterization and interpretation of invertebrate assemblages. Journal of the North American Benthological Society. 2007;26:286–307. doi: 10.1899/0887-3593(2007)26[286:ATEOTC]2.0.CO;2. [DOI] [Google Scholar]
  • Dale & Beyeler (2001).Dale VH, Beyeler SC. Challenges in the development and use of ecological indicators. Ecological Indicators. 2001;1:3–10. doi: 10.1016/S1470-160X(01)00003-6. [DOI] [Google Scholar]
  • Darling et al. (2013).Darling ES, Shiffman D, Côte IM, Drew JA. The role of twitter in the life cycle of a scientific publication. Ideas in Ecology and Evolution. 2013;6:32–43. [Google Scholar]
  • Davies & Jackson (2006).Davies SP, Jackson SK. The biological condition gradient: a descriptive model for interpreting change in aquatic ecosystems. Ecological Applications. 2006;16:1251–1266. doi: 10.1890/1051-0761(2006)016[1251:TBCGAD]2.0.CO;2. [DOI] [PubMed] [Google Scholar]
  • De Caceres & Legendre (2009).De Caceres M, Legendre P. Ecology. 2009. Associations between species and groups of sites: indices and statistical inference. [DOI] [PubMed] [Google Scholar]
  • Deiner et al. (2017).Deiner K, Bik HM, Mächler E, Seymour M, Lacoursiére-Roussel A, Altermatt F, Creer S, Bista I, Lodge DM, de Vere N, Pfrender ME, Bernatchez L. Environmental DNA metabarcoding: transforming how we survey animal and plant communities. Molecular Ecology. 2017;26:5872–5895. doi: 10.1111/mec.14350. [DOI] [PubMed] [Google Scholar]
  • Dickersin & Rennie (2003).Dickersin K, Rennie D. Registering clinical trials. JAMA. 2003;290(4):516–523. doi: 10.1001/jama.290.4.516. [DOI] [PubMed] [Google Scholar]
  • Elbrecht et al. (2017).Elbrecht V, Vamos EE, Meissner K, Aroviita J, Leese F. Assessing strengths and weaknesses of dna metabarcoding-based macroinvertebrate identification for routine stream monitoring. Methods in Ecology and Evolution. 2017;8:1265–1275. doi: 10.1111/2041-210X.12789. [DOI] [Google Scholar]
  • Essl et al. (2020).Essl F, Courchamp F, Dullinger S, Jeschke JM, Schindler S. Make open access publishing fair and transparent! BioScience. 2020;70:201–204. doi: 10.1093/biosci/biaa004. [DOI] [Google Scholar]
  • Fetscher et al. (2013).Fetscher AE, Stancheva R, Kociolek JP, Sheath RG, Stein ED, Mazor RD, Ode PR, Busse LB. Development and comparison of stream indices of biotic integrity using diatoms vs. Non-diatom algae vs. A combination. Journal of Applied Phycology. 2013;26:433–450. [Google Scholar]
  • Fore & Grafe (2002).Fore LS, Grafe C. Using diatoms to assess the biological condition of large rivers in Idaho (U.S.A.) Freshwater Biology. 2002;47:2015–2037. doi: 10.1046/j.1365-2427.2002.00948.x. [DOI] [Google Scholar]
  • Franco, Malhotra & Simonovits (2014).Franco A, Malhotra N, Simonovits G. Publication bias in the social sciences: unlocking the file drawer. Science. 2014;345:1502–1505. doi: 10.1126/science.1255484. [DOI] [PubMed] [Google Scholar]
  • Grand et al. (2012).Grand A, Wilkinson C, Bultitude K, Winfield AFT. Open science: A new “trust technology”? Science Communication. 2012;34:679–689. doi: 10.1177/1075547012443021. [DOI] [Google Scholar]
  • Hampton et al. (2015).Hampton SE, Anderson SS, Bagby SC, Gries C, Han X, Hart EM, Jones MB, Lenhardt WC, MacDonald A, Michener WK, Mudge J, Pourmokhtarian A, Schildhauer MP, Woo KH, Zimmerman N. The tao of open science for ecology. Ecosphere. 2015;6:1–13. [Google Scholar]
  • Hampton et al. (2017).Hampton SE, Jones MB, Wasser LA, Schildhauer MP, Supp SR, Brun J, Hernandez RR, Boettiger C, Collins SL, Gross LJ, Fernández DS, Budden A, White EP, Teal TK, Labou SG, Aukema JE. Skills and knowledge for data-intensive environmental research. Bioscience. 2017;67:546–557. doi: 10.1093/biosci/bix025. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • Hampton et al. (2016).Hampton SE, Strasser CA, Tewksbury JJ, Gram WK, Budden AE, Batcheller AL, Duke CS, Porter JH. Big data and the future of ecology. Frontiers in Ecology and the Environment. 2016;11:156–162. [Google Scholar]
  • Hawkins et al. (2000a).Hawkins CP, Norris RH, Gerritsen J, Hughes RM, Jackson SK, Johnson RK, Stevenson RJ. Evaluation of the use of landscape classifications for the prediction of freshwater biota: synthesis and recommendations. Journal of the North American Benthological Society. 2000a;19:541–556. doi: 10.2307/1468113. [DOI] [Google Scholar]
  • Hawkins et al. (2000b).Hawkins CP, Norris RH, Hogue JN, Feminella JW. Development and evaluation of predictive models for measuring the biological integrity of streams. Ecological Applications. 2000b;10:1456–1477. doi: 10.1890/1051-0761(2000)010[1456:DAEOPM]2.0.CO;2. [DOI] [Google Scholar]
  • Hawkins, Olson & Hill (2010).Hawkins CP, Olson JR, Hill RA. The reference condition: predicting benchmarks for ecological and water-quality assessments. Journal of the North American Benthological Society. 2010;29:312–343. doi: 10.1899/09-092.1. [DOI] [Google Scholar]
  • Hering et al. (2010).Hering D, Borja A, Carstensen J, Carvalho L, Elliott M, Field CK, Heiskanen AS, Johnson RK, Moe J, Pont D, Solheim AL, de Bund W. The European water framework directive at the age of 10: a critical review of the achievements with recommendations for the future. Science of the Total Environment. 2010;408:4007–4019. doi: 10.1016/j.scitotenv.2010.05.031. [DOI] [PubMed] [Google Scholar]
  • Hering et al. (2018).Hering D, Borja A, Jones JI, Pont D, Boets P, Bouchez A, Bruce K, Drakare S, Hänfling B, Kahlert M, Leese F, Meissner K, Mergen P, Reyjol Y, Segurado P, Vogler A, Kelly M. Implementation options for DNA-based identification into ecological status assessment under the European Water Framework Directive. Water Research. 2018;138:192–205. doi: 10.1016/j.watres.2018.03.003. [DOI] [PubMed] [Google Scholar]
  • Hijmans (2019).Hijmans RJ. Raster: Geographic data analysis and modeling. https://CRAN.R-project.org/package=raster 2019
  • Hill et al. (2016).Hill RA, Weber MH, Leibowitz SG, Olsen AR, Thornbrugh DJ. The Stream-Catchment (StreamCat) dataset: a database of watershed metrics for the conterminous United States. Journal of the American Water Resources Association. 2016;52:120–128. doi: 10.1111/1752-1688.12372. [DOI] [Google Scholar]
  • Horsburgh et al. (2016).Horsburgh JS, Morsy MM, Castronova AM, Goodall JL, Gan T, Yi H, Stealey MJ, Tarboton DG. Hydroshare: sharing diverse environmental data types and models as social objects with application to the hydrology domain. Journal of the American Water Resources Association. 2016;52:873–889. doi: 10.1111/1752-1688.12363. [DOI] [Google Scholar]
  • Hsu et al. (2017).Hsu L, Mayorga E, Horsburgh JS, Carter MR, Lehnert KA, Brantley SL. Enhancing interoperability and capabilities of earth science data using the Observations Data Model 2 (ODM2) Data Science Journal. 2017;16:1–16. doi: 10.5334/dsj-2017-001. [DOI] [Google Scholar]
  • Idaszak et al. (2017).Idaszak R, Tarboton DG, Yi H, Chistopherson L, Stealey MJ, Miles B, Dash P, Couch A, Spealman C, Ames DP, Horsburgh JS. HydroShare - a case study of the application of modern software engineering to a large distributed federally-funded scientific software development project. In: Carver JC, Hong NPC, Thiruvathukal GK, editors. Milton Park: Taylor & Francis Group; 2017. pp. 217–233. [Google Scholar]
  • Ihle et al. (2017).Ihle M, Winney IS, Krystalli A, Croucher M. Striving for transparent and credible research: practical guidelines for behavioral ecologists. Behavioral Ecology. 2017;28:348–354. doi: 10.1093/beheco/arx003. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • Karr et al. (1986).Karr JR, Fausch KD, Angermeier PL, Yant PR, Schlosser IJ. Assessing biological integrity in running waters: a method and its rationale. Illinois Natural History Survey; Champaign: 1986. [Google Scholar]
  • Kelly et al. (2016).Kelly MG, Birk S, Willby NJ, Denys L, Drakare S, Kahlert M, Karjalainen SM, Marchetto A, Pitt JA, Urbanič G, Poikane S. Redundancy in the ecological assessment of lakes: are phytoplankton, macrophytes and phytobenthos all necessary? Science of the Total Environment. 2016;15:594–602. doi: 10.1016/j.scitotenv.2016.02.024. [DOI] [PubMed] [Google Scholar]
  • Kerans & Karr (1994).Kerans BL, Karr JR. A benthic index of biotic integrity (B-IBI) for rivers of the Tennessee Valley. Ecological Applications. 1994;4:768–785. doi: 10.2307/1942007. [DOI] [Google Scholar]
  • King et al. (2011).King RS, Baker ME, Kazyak PF, Weller DE. How novel is too novel? Stream community thresholds at exceptionally low levels of catchment urbanization. Ecological Applications. 2011;21:1659–1678. doi: 10.1890/10-1357.1. [DOI] [PubMed] [Google Scholar]
  • Kluyver et al. (2016).Kluyver T, Ragan-Kelley B, Pérez F, Granger B, Bussonnier M, Frederic J, Kelley K, Hamrick J, Grout J, Corlay S, Ivanov P, Avila D, Abdalla S, Willing C. Jupyter notebooks –a publishing format for reproducible computational workflows. Amsterdam: IOS Press; 2016. [Google Scholar]
  • Kuehne et al. (2017).Kuehne LM, Olden JD, Strecker AL, Lawler JJ, Theobald DM. Past, present, and future of ecological integrity assessment for fresh waters. Frontiers in Ecology and the Environment. 2017;15:197–205. doi: 10.1002/fee.1483. [DOI] [Google Scholar]
  • Kuehne, Strecker & Olden (2019).Kuehne LM, Strecker AL, Olden JD. Knowledge exchange and social capital for freshwater ecosystem assessments. BioScience. 2019;70:174–183. [Google Scholar]
  • Laan & Hawkins (2014).Laan JJV, Hawkins CP. Enhancing the performance and interpretation of freshwater biological indices: an application in arid zone streams. Ecological Indicators. 2014;36:470–482. doi: 10.1016/j.ecolind.2013.09.006. [DOI] [Google Scholar]
  • Lai et al. (2019).Lai J, Lortie CJ, Muenchen RA, Yang J, Ma K. Evaluating the popularity of R in ecology. Ecosphere. 2019;10:e02567 [Google Scholar]
  • Landman & Glantz (2009).Landman A, Glantz SA. Tobacco industry efforts to undermine policy-relevant research. American Journal of Public Health. 2009;99:45–58. doi: 10.2105/AJPH.2007.130740. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • Lenat & Resh (2001).Lenat DR, Resh VH. Taxonomy and stream ecology—the benefits of genus- and species-level identifications. Journal of the North American Benthological Society. 2001;20:287–298. doi: 10.2307/1468323. [DOI] [Google Scholar]
  • Lewandowsky & Bishop (2016).Lewandowsky S, Bishop D. Research integrity: don’t let transparency damage science. Nature. 2016;529:459–461. doi: 10.1038/529459a. [DOI] [PubMed] [Google Scholar]
  • Liaw & Wiener (2002).Liaw A, Wiener M. Classification and regression by randomForest. R News. 2002;2:18–22. [Google Scholar]
  • Liu et al. (2008).Liu Y, Gupta H, Springer E, Wagener T. Linking science with environmental decision making: experiences from an integrated modeling approach to supporting sustainable water resources management. Environmental Modelling & Software. 2008;23:846–858. doi: 10.1016/j.envsoft.2007.10.007. [DOI] [Google Scholar]
  • Lortie (2014).Lortie CJ. Formalized synthesis opportunities for ecology: systematic reviews and meta-analyses. OIKOS. 2014;123:897–902. doi: 10.1111/j.1600-0706.2013.00970.x. [DOI] [Google Scholar]
  • Lowndes et al. (2017).Lowndes JSS, Best BD, Scarborough C, Afflerbach JC, Frazier MR, O’Hara CC, Jiang N, Halpern BS. Our path to better science in less time using open data science tools. Nature Ecology & Evolution. 2017;1:1–7. doi: 10.1038/s41559-016-0001. [DOI] [PubMed] [Google Scholar]
  • Makel, Plucker & Hegarty (2012).Makel MC, Plucker JA, Hegarty B. Replications in psychology research: How often do they really occur? Perspectives on Psychological Science. 2012;7:537–542. doi: 10.1177/1745691612460688. [DOI] [PubMed] [Google Scholar]
  • Marwick et al. (2016).Marwick B, d’Alpoim Guedes J, Barton CM, Bates LA, Baxter M, Bevan A, Bollwerk EA, Bocinsky RK, Brughmans T, Carter AK, Conrad C, Contreras DA, Costa S, Crema ER, Daggett A, Davies B, Drake BL, Dye TS, France P, Fullagar R, Giusti D, Graham S, Harris MD, Hawks J, Heath S, Huffer D, Kansa EC, Kansa SW, Madsen ME, Melcher J, Negre J, Neiman FD, Opitz R, Orton DC, Przystupa P, Raviele M, Riel-Salvatore J, Riris P, Romanowska I, Strupler N, Ullah II, Vlack HGV, Watrall EC, Webster C, Wells J, Winters J, Wren CD. Open science in archaeology. SAA Archaeological Record. 2016;17:8–14. [Google Scholar]
  • Mazor et al. (2016).Mazor RD, Rehn AC, Ode PR, Engeln M, Schiff KC, Stein ED, Gillett DJ, Herbst DB, Hawkins CP. Bioassessment in complex environments: designing an index for consistent meaning in different settings. Freshwater Science. 2016;35:249–271. doi: 10.1086/684130. [DOI] [Google Scholar]
  • Mazor et al. (2014).Mazor RD, Stein ED, Ode PR, Schiff K. Integrating intermittent streams into watershed assessments: applicability of an index of biotic integrity. Freshwater Science. 2014;35:459–474. [Google Scholar]
  • McKay et al. (2012).McKay L, Bondelid T, Dewald T, Johnston J, Moore R, Reah A. NHDPlus version 2: user guide. https://nctc.fws.gov/courses/references/tutorials/geospatial/CSP7306/Readings/NHDPlusV2_User_Guide.pdf 2012
  • Michener et al. (1997).Michener WK, Brunt JW, Helly JJ, Kirchner TB, Stafford SG. Nongeospatial metadata for the ecological sciences. Ecological Applications. 1997;7:330–342. doi: 10.1890/1051-0761(1997)007[0330:NMFTES]2.0.CO;2. [DOI] [Google Scholar]
  • Mitchell (2005).Mitchell B. Integrated water resource management, institutional arrangements, and land use planning. Environmental Planning A: Economy and Space. 2005;37:1335–1352. doi: 10.1068/a37224. [DOI] [Google Scholar]
  • Munafó et al. (2017).Munafó MR, Nosek BA, Bishop DVM, Button KS, Chambers CD, du Sert NP, Simonsohn U, Wagenmakers EJ, Ware JJ, Loannidis JPA. A manifesto for reproducible science. Nature Human Behavior. 2017;1:0021. doi: 10.1038/s41562-016-0021. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • Nichols et al. (2016).Nichols SJ, Barmuta LA, Chessman BC, Davies PE, Dyer FJ, Harrison ET, Hawkins CP, Jones I, Kefford BJ, Linke S, Marchant R, Metzeling L, Moon K, Ogden R, Peat M, Reynoldson TB, Thompson RM. The imperative need for nationally coordinated bioassessment of rivers and streams. Marine and Freshwater Research. 2016;68:599–613. [Google Scholar]
  • Nosek et al. (2018).Nosek BA, Ebersole CR, DeHaven AC, Mellor DT. The preregistration revolution. Proceedings of the National Academy of Sciences. 2018;115:2600–2606. doi: 10.1073/pnas.1708274114. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • Ode et al. (2016).Ode PR, Rehn AC, Mazor RD, Schiff KC, Stein ED, May JT, Brown LR, Herbst DB, Gillett D, Lunde K, Hawkins CP. Evaluating the adequacy of a reference-site pool for ecological assessments in environmentally complex regions. Freshwater Science. 2016;35:237–248. doi: 10.1086/684003. [DOI] [Google Scholar]
  • Oksanen et al. (2018).Oksanen J, Blanchet FG, Friendly M, Kindt R, Legendre P, McGlinn D, Minchin PR, O’Hara RB, Simpson GL, Solymos P, Stevens MHH, Szoecs E, Wagner H. Vegan: community ecology package. https://CRAN.R-project.org/package=vegan 2018
  • Pebesma (2018).Pebesma E. Sf: Simple features for r. https://CRAN.R-project.org/package=sf 2018
  • Pilkington (2016).Pilkington M. Blockchain technology: principles and applications. In: Olleros FX, Zhegu M, editors. Research handbook on digital transformations. Cheltenham: Edward Elgar Publishing; 2016. pp. 225–253. [Google Scholar]
  • Pinheiro et al. (2018).Pinheiro J, Bates D, DebRoy S, Sarkar D, R Core Team nlme: Linear and nonlinear mixed effects models. https://CRAN.R-project.org/package=nlme 2018
  • Pohjola & Tuomisto (2011).Pohjola MV, Tuomisto JT. Openness in participation, assessment, and policy making upon issues of environment and environmental health: a review of literature and recent project results. Environmental Health. 2011;10:1–13. doi: 10.1186/1476-069X-10-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • Poikane et al. (2014).Poikane S, Zampoukas N, Borja A, Davies SP, Van de Bund W, Birk S. Intercalibration of aquatic ecological assessment methods in the European Union. Environmental Science & Policy. 2014;44:237–246. doi: 10.1016/j.envsci.2014.08.006. [DOI] [Google Scholar]
  • Powers & Hampton (2019).Powers SM, Hampton SE. Open science, reproducibility, and transparency in ecology. Ecological Applications. 2019;29:e01822. doi: 10.1002/eap.1822. [DOI] [PubMed] [Google Scholar]
  • RDCT (2020).RDCT (R Development Core Team) R: a language and environment for statistical computing, v4.0.2. R Foundation for Statistical Computing; Vienna: 2020. [Google Scholar]
  • Rehn, Mazor & Ode (2018).Rehn AC, Mazor RD, Ode PR. Sacramento: California Water Boards, Surface Water Ambient Monitoring Program, California Department of Fish; Wildlife, Southern California Coastal Water Research ProjectAn index to measure the quality of physical habitat in California wadeable streams. 2018
  • Slater et al. (2019).Slater LJ, Thirel G, Harrigan S, Delaigue O, Hurley A, Khouakhi A, Prosdocimi I, Vitolo C, Smith K. Using R in hydrology: a review of recent developments and future directions. Hydrology and Earth System Sciences. 2019;23:2939–2963. doi: 10.5194/hess-23-2939-2019. [DOI] [Google Scholar]
  • Stein et al. (2009).Stein ED, Brinson M, Rains MC, Kleindl W, Hauer FR. Wetland assessment alphabet soup: How to choose (or not choose) the right assessment method. Society of Wetland Scientists Bulletin. 2009;26:20–24. [Google Scholar]
  • Stevens et al. (2018).Stevens SLR, Kuzak M, Martinez C, Moser A, Bleeker P, Galland M. Building a local community of practice in scientific programming for life scientists. PLOS Biology. 2018;16:e2005561. doi: 10.1371/journal.pbio.2005561. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • Stoddard et al. (2006).Stoddard JL, Larsen DP, Hawkins CP, Johnson RK, Norris RH. Setting expectations for the ecological condition of streams: the concept of reference condition. Ecological Applications. 2006;16:1267–1276. doi: 10.1890/1051-0761(2006)016[1267:SEFTEC]2.0.CO;2. [DOI] [PubMed] [Google Scholar]
  • Taylor et al. (2018).Taylor JM, Back JA, Brooks BW, King RS. Spatial, temporal and experimental: three study design cornerstones for establishing defensible numeric criteria in freshwater ecosystems. Journal of Applied Ecology. 2018;55:2114–2123. doi: 10.1111/1365-2664.13150. [DOI] [Google Scholar]
  • Taylor et al. (2014).Taylor JM, King RS, Pease AA, Winemiller KO. Nonlinear response of stream ecosystem structure to low-level phosphorus enrichment. Freshwater Biology. 2014;59:969–984. doi: 10.1111/fwb.12320. [DOI] [Google Scholar]
  • Touchon & McCoy (2016).Touchon JC, McCoy MW. The mismatch between current statistical practice and doctoral training in ecology. Ecosphere. 2016;7:e01394. doi: 10.1002/ecs2.1394. [DOI] [Google Scholar]
  • Van Oudenhoven, Schröter & De Groot (2016).Van Oudenhoven APE, Schröter M, De Groot R. Open access to science on ecosystem services and biodiversity. International Journal of Biodiversity Science, Ecosystem Services & Management. 2016;12:239–241. [Google Scholar]
  • White et al. (2014).White BP, Pilgrim EM, Boykin LM, Stein ED, Mazor RD. Comparison of four species-delimitation methods applied to a DNA barcode data set of insect larvae for use in routine bioassessment. Freshwater Science. 2014;33:338–348. doi: 10.1086/674982. [DOI] [Google Scholar]
  • Wickham (2009).Wickham H. Ggplot2: elegant graphics for data analysis. Springer-Verlag; New York: 2009. [Google Scholar]
  • Wickham (2015).Wickham H. R packages. http://r-pkgs.had.co.nz/ 2015
  • Wickham (2017).Wickham H. Tidyverse: easily install and load the ‘tidyverse’. https://CRAN.R-project.org/package=tidyverse 2017
  • Wickham, Hester & Chang (2018).Wickham H, Hester J, Chang W. Devtools: tools to make developing r packages easier. https://CRAN.R-project.org/package=devtools 2018
  • Wilkinson (2005).Wilkinson L. The grammar of graphics. Second edition. Statistics; Computing, Springer; New York: 2005. p. 691. [Google Scholar]
  • Woelfle, Olliaro & Todd (2011).Woelfle M, Olliaro P, Todd MH. Open science is a research accelerator. Nature Chemistry. 2011;3:745–748. doi: 10.1038/nchem.1149. [DOI] [PubMed] [Google Scholar]
  • Wood (2017).Wood SN. Generalized additive models: an introduction with r. 2nd edition. Chapman; Hall, CRC Press; London: 2017. p. 476. [Google Scholar]
  • Xie (2015).Xie Y. Dynamic documents with R and knitr. 2nd edition Chapman; Hall/CRC; Boca Raton: 2015. [Google Scholar]
  • Yenni et al. (2019).Yenni GM, Christensen EM, Bledsoe EK, Supp SR, Diaz RM, White EP, Ernest SKM. Developing a modern data workflow for regularly updated data. PLOS Biology. 2019;17:e3000125. doi: 10.1371/journal.pbio.3000125. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • Yuan (2004).Yuan LL. Assigning macroinvertebrate tolerance classifications using generalised additive models. Freshwater Biology. 2004;49:662–677. doi: 10.1111/j.1365-2427.2004.01206.x. [DOI] [Google Scholar]
  • Zipper et al. (2019).Zipper SC, Whitney KSS, Deines JM, Befus KM, Bhatia U, Albers SJ, Beecher J, Brelsford C, Garcia M, Gleeson T, O’Donnell F, Resnik D, Schlager E. Balancing open science and data privacy in the water sciences. Water Resources Research. 2019;55:1–10. doi: 10.1002/wrcr.23423. [DOI] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

The following information was supplied regarding data availability:

All source materials for reproducing the paper are available in GitHub: https://github.com/fawda123/bioassess_opensci.


Articles from PeerJ are provided here courtesy of PeerJ, Inc

RESOURCES