Skip to main content
Applications in Plant Sciences logoLink to Applications in Plant Sciences
. 2020 Jul 1;8(6):e11352. doi: 10.1002/aps3.11352

Generating segmentation masks of herbarium specimens and a data set for training segmentation models using deep learning

Alexander E White 1,2, Rebecca B Dikow 1,, Makinnon Baugh 3, Abigail Jenkins 3, Paul B Frandsen 1,3
PMCID: PMC7328659  PMID: 32626607

Abstract

Premise

Digitized images of herbarium specimens are highly diverse with many potential sources of visual noise and bias. The systematic removal of noise and minimization of bias must be achieved in order to generate biological insights based on the plants rather than the digitization and mounting practices involved. Here, we develop a workflow and data set of high‐resolution image masks to segment plant tissues in herbarium specimen images and remove background pixels using deep learning.

Methods and Results

We generated 400 curated, high‐resolution masks of ferns using a combination of automatic and manual tools for image manipulation. We used those images to train a U‐Net‐style deep learning model for image segmentation, achieving a final Sørensen–Dice coefficient of 0.96. The resulting model can automatically, efficiently, and accurately segment massive data sets of digitized herbarium specimens, particularly for ferns.

Conclusions

The application of deep learning in herbarium sciences requires transparent and systematic protocols for generating training data so that these labor‐intensive resources can be generalized to other deep learning applications. Segmentation ground‐truth masks are hard‐won data, and we share these data and the model openly in the hopes of furthering model training and transfer learning opportunities for broader herbarium applications.

Keywords: deep learning, digitized herbarium specimens, ferns, machine learning, semantic segmentation, U‐Net


Efforts to digitize natural history collections have resulted in large, open data sets that, when linked with collection data, offer an unprecedented opportunity to evaluate transformative research questions that link objects through space and time. The tens of millions of openly available digitized herbarium specimens allow botanists to evaluate ecological and evolutionary questions on a global scale, integrating across centuries of botanical collecting and piecing together collections with spatially biased holdings to generate biome‐level comparisons of biological patterns (Davis et al., 2015; Pearse et al., 2017; Park and Mazer, 2018). Collating millions of herbarium specimens into a single analysis, however, necessarily involves aggregating noise, error, and bias from thousands of different sampling, mounting, and digitization protocols, given that herbarium sheets are digitized in independent efforts in the approximately ~3000 herbaria worldwide (Thiers, 2018). Such bias and noise may influence not only the analysis of digitized collection records (e.g., Daru et al., 2018), but also analyses of specimen images, a growing source of data for botanical questions ranging from phenology (e.g., Willis et al., 2017; Lorieul et al., 2019) to morphology (e.g., Kavanagh et al., 2011; Burns et al., 2012; Easlon and Bloom, 2014) to global change biology (e.g., Meineke et al., 2018; Heberling et al., 2019). To move beyond the important efforts of quantifying sources of bias in the specimen records, we must therefore begin to use statistical methods that allow for the identification and elimination of sources of bias in image data to leverage images of herbarium specimens to their full potential.

Machine learning and its subfield of deep learning are particularly useful for the analysis of specimen images (Unger et al., 2016), as these types of models can be trained to identify and ignore sources of image variation and noise. For example, deep learning models in the form of convolutional neural networks (CNNs; LeCun et al., 2015) are well suited to classifying objects in images irrespective of the position or orientation of those objects. Although a variety of object orientations or contexts might pose a challenge to more traditional quantitative methods of classification, these sources of noise do not prevent deep learning models from achieving high accuracy for object identification. The ability of deep learning models to maintain both high classification accuracy and wide generality does not, however, preclude them from identifying a given image class (e.g., taxonomic identity) based on information that is otherwise non‐biological. Indeed, there are a number of visual cues that may allow a specimen to be identified based on the herbarium from which it originated rather than the features of the plant itself. Other biologically uninformative visual information may include rulers, a color bar or palette, collection‐specific barcodes, stamps, collection identifiers, text that might be included as standard in a given collection, or even accumulated sources of debris (e.g., mercuric chloride staining; see Schuettpelz et al., 2017). Although most of these visual cues may be easily interpreted as noise, which neural networks are well suited to ignore, the critical concern is unknown cases where these cues are unique, rare, and singularly associated with a specific class that the model is built to identify. These cues may undermine the performance of a deep learning model on a novel data set that lacks such cues. Systematic and scalable methods are needed to remove such cues from analyses that rely on digitized herbarium specimens as the primary source of data, particularly for deep learning data sets where the scale of the sample group often prohibits the manual inspection of each image.

One promising solution is to identify biologically relevant image pixels (i.e., plant tissues) a priori, eliminating noisy and biased visual information from all other parts of the herbarium specimen image ahead of subsequent analyses. Deep learning models can be trained to label image pixels based on their content (in this case, RGB values). Labeling image pixels in this way is referred to as semantic segmentation, and has been applied in other botanical applications, for example in automated agriculture (Milioto et al., 2018). Novel deep learning model architectures are also being developed to improve focal object attention during classification (e.g., attention‐based CNNs, Ba et al., 2014; Simonyan and Zisserman, 2014; Ren et al., 2015) and applied in the medical field (e.g., Li et al., 2019); however, there are also a number of potential neural network applications for these models in herbarium science beyond classification. In such applications (e.g., automated identification and measurement of specific plant tissues; Lorieul et al., 2019), the semantic segmentation of herbarium specimens may be critical for generating meaningful biological insights.

The primary and perhaps most critical step in developing an accurate deep learning model for semantic segmentation is to generate a high‐resolution data set of images with their associated masks, images of identical resolution that define the identity of each pixel in the original image as belonging to an a priori designated set of pixel identities or classes (e.g., plant tissue, label, color bar). These data (original images paired with their masks) are used to train the deep learning model to generate accurate pixel classifications by comparing the model predictions for each pixel against the pixel class defined in the mask (i.e., the ground truth). A more detailed description of the learning process is beyond the scope of this work, but see Garcia‐Garcia et al. (2017) for a general review of semantic segmentation and its applications in deep learning. Although image masks for herbarium specimens can be as simple as a binary image identifying two classes, one for all pixels containing plant tissues and another for all other visual information (background), herbarium specimens contain such a wide diversity of plant sizes and shapes that drawing boundaries along the edges of the specimens, particularly around complex leaf structures, is extremely labor intensive and nearly impossible on a large scale. Generating masks of complex plant structures at high resolution is not only labor intensive but also particularly difficult using polygon‐based annotation approaches such as those offered in ImageJ (Schneider et al., 2012; Rueden et al., 2017). In this paper, we present a systematic workflow for generating high‐quality image masks of digitized herbarium specimens for use with deep learning segmentation models. We describe our efforts to train and validate such a model using a Python‐based deep learning framework and share an open repository where the original images, ground‐truth masks, and the trained deep learning model can be accessed for future use outside the application we detail below.

Potential uses for the image segmentation of digitized herbarium images include the detection of flowers, the quantification of fruits, and the estimation of the intensity of disease and the extent of herbivory (e.g., Meineke and Davies, 2019); the data and model we present could be leveraged for those applications as well. Similar uses exist across different fields where machine learning has become an essential tool; for example, these types of analyses are vital in the medical field, where precision in the determination of tissue types and pathologies in medical images is essential for accurate diagnosis and successful treatment decisions made by medical professionals (e.g., Ciresan et al., 2012; Roth et al., 2015; Moeskops et al., 2016; Zhuang et al., 2019). There are other industrial applications of this technology beyond medicine, one of which, agriculture, is more closely related to the work we present here. For example, whereas high‐throughput genotyping has been used in plant breeding applications for decades, the nascent field of high‐throughput phenotyping is focused on generating reliable genotype–phenotype associations (Choudhury et al., 2019). The accurate evaluation of changes in phenotype requires models that result in well‐resolved measurements of phenotypic change (Singh et al., 2016).

Our workflow draws upon best practices and recent advances in computer vision and deep learning, making use of openly available Python libraries for image manipulation (OpenCV, Bradski and Kaehler, 2008; PlantCV, Gehan et al., 2017) and machine learning (PyTorch, Paszke et al., 2017; fastai, Howard et al., 2018). The deep learning model architecture we employ (U‐Net; Ronneberger et al., 2015) was originally developed in a medical context, but is now well known for its efficient and accurate performance in general image segmentation tasks. We combine these practices in one protocol to allow more systematic improvement and transparency regarding training data inputs in machine learning applications in herbarium science. Critically, to motivate model sharing and facilitate transfer learning applications within the herbarium science community, we share the trained model and these data, including both high‐resolution images and image masks, on Figshare (see Data Availability).

MATERIALS AND RESULTS

Data

Our goal was to generate a segmentation model for plant tissues with complex margins, yet we felt that some measure of taxonomic specificity would increase the model’s accuracy and utility. Ferns are known to exhibit an incredible diversity of leaf forms (see Vasco et al. [2013] for a comprehensive review of fern leaf morphology) and, in the absence of fruits, flowers, and woody material, fern specimens in herbaria largely reflect leaf material mounted on paper. Recognizing both the broad diversity and complexity of fern leaf margins as well as their relatively simple representation across specimens within herbaria, we limited our data set to only include images of fern specimens. We obtained digital images (high‐resolution JPEG [.jpg] files, ca. 9100 × 6800 pixels) of all such specimens housed in the U.S. National Herbarium at the National Museum of Natural History (Washington, D.C., USA). The images were acquired from the Smithsonian Institution Digital Asset Management System and transferred to the Smithsonian High‐Performance Computing Cluster for analysis. To curate these images into a manageable data set for mask generation, we chose 400 random images and subjectively hand‐curated our image set to reflect the broad diversity of leaf shapes present in ferns as much as possible (Fig. 1A); replacement images were chosen at random until the set of images subjectively appeared to reflect the wide morphological diversity of ferns. Images without plant tissue (e.g., specimens housed in envelopes and affixed to sheets) were removed and replaced with another random specimen containing visible tissues. In total, the 400 images we curated include specimens from 11 orders, 30 families, 99 genera, and 308 species (specimen metadata is included, see Data Availability).

FIGURE 1.

FIGURE 1

Herbarium sheets and associated masks made available in this study. (A) Four example digitized herbarium sheets from the U.S. National Herbarium at the National Museum of Natural History (Washington, D.C., USA). (B) The same four sheets shown as high‐resolution masks. A total of 400 masks were generated using the methods described in the text and were used to train a deep neural net to automatically segment plant tissues from herbarium specimens. Species names for each image clockwise from top left: Rumohra adiantiformis (G. Forst.) Ching, Thelypteris kunthii (Desv.) C. V. Morton (synonym Christella kunthii), Asplenium peruvianum var. insulare (C. V. Morton) D. D. Palmer, Thelypteris palustris Schott.

Protocol for generating image masks

We developed a workflow to generate ground‐truthed herbarium image masks for a simple binary segmentation task in which all pixels are labeled as either plant pixels or background pixels (Fig. 1B). This protocol combines a thresholding method of automatic image segmentation (Otsu, 1979) with manual postprocessing and mask editing to generate high‐resolution and high‐quality image masks (Fig. 2). We describe each component of our workflow below.

FIGURE 2.

FIGURE 2

Workflow outlining the automatic and manual steps in generating the image masks and training the U‐Net. High‐resolution JPEG (.jpg) files were exported from the Smithsonian Digital Asset Management System to the High‐Performance Computing Cluster where we ran the segmentation Python code. Outputs from this step were edited in Adobe Photoshop to remove label and color palette before running the postprocessing code (binarize and blur tools) that produced the final ground‐truth masks. These ground‐truth masks were then used as training data for the U‐Net model.

First, we import a grayscale version of each original high‐resolution image into Python (van Rossum and Drake, 2009) using the imread method in OpenCV (Bradski and Kaehler, 2008). We then use Otsu’s binarization method (Otsu, 1979) as implemented in OpenCV to estimate image thresholds. In short, Otsu’s method searches the distribution of pixel values to estimate a threshold that minimizes intraclass variance. We then assign a binary value (either black [0 in byte image values] or white [255]) to each pixel according to the Otsu threshold to create both foreground and background images. The foreground image is the inverse of the background image. We export both the foreground and background images with filenames corresponding to the original image. We automatically generated these preliminary masks in Python version 3.7 using OpenCV version 4.0.1. The code is available at https://github.com/sidatasciencelab/fern_segmentation.

The ideal image mask would include a single pixel value for plant tissue and another pixel value for the rest of the herbarium sheet, thus allowing the extraction of plant‐only elements for downstream applications. Although the segmentation process described above worked well to include all plant materials in one grouping, each mask also included herbarium elements, such as label data and the color palette (see top middle box, Fig. 2). Because removing these elements automatically is particularly challenging, we instead developed a manual process for their removal. We chose the foreground images (those with white pixels assigned to the plant tissue) and manually edited them in Adobe Photoshop CC 2018 (Adobe Inc., San Jose, California, USA). We used the Brush Tool and the Rectangle Tool to adjust the pixels that should have been assigned a black (0) pixel value during segmentation but were otherwise mislabeled. Due to the behavior of these tools in Photoshop, some of the edge pixels were assigned intermediate gray values, yet pixels needed to be assigned binary pixel values according to our objective. Thus, after we sufficiently edited the extraneous herbarium elements from the images in Photoshop, we binarized the mask images (using the binarize function in OpenCV) to ensure that the pixel values were either 0 or 255.

As a final step in preparing the images for our data set, we utilized the blur method contained within the PlantCV library (Gehan et al., 2017) to remove any remaining disparities (e.g., there were images where dust particles were labeled as plant material and blur uses pixel neighborhood values to smooth away these disparities). For all images, we used three different parameter values and chose the “best” mask by visually inspecting them. This was necessary because the plants vary in size, age, and overall condition (e.g., amount of debris).

Training a deep learning model for segmentation

We trained a PyTorch (version 1.1.0, Paszke et al., 2017) deep learning model for binary image segmentation of herbarium images using fastai (version 1.0.55, Howard et al., 2018) in Python 3.7. Following an emerging standard of best practices for data set preparation and image transformations (He et al., 2018), we trained a U‐Net style neural network using 80% (n = 320) of our original images paired with ground‐truthed masks prepared using the protocol above. This resulted in approximately 21 million pixels with associated class labels for training this model. The goal of such training is to expose the model to a wide diversity of pixel values and contexts paired with the associated pixel class identity (plant or background). All images and associated masks were resized to 256 × 256 pixels to maximize the downstream training efficiency. Our model therefore produces predicted image masks of 256 × 256 pixels regardless of the size of the image input (Fig. 3). The square output predictions crop image inputs if they are rectangular (see Fig. 3C).

FIGURE 3.

FIGURE 3

A comparison of high‐resolution original images, ground‐truth masks, and U‐Net‐predicted mask outputs. (A) Two example original images. (B) Ground‐truth masks. (C) Mask outputs predicted by U‐Net (Sørensen–Dice coefficient = 0.95). Note that the predicted masks are all resized to 256 × 256 pixels to maximize downstream model training efficiency regardless of image input size. The square output predictions crop rectangular inputs. Species names from top: Callistopteris apiifolia (Presl) Copel., Ceradenia capillaris (Desv.) L. E. Bishop.

We transformed our data randomly during the training process using image augmentation, as is standard practice to maximize the generality of deep learning models. Transformations included flipping images horizontally, rotating (maximum rotation 10 degrees), zooming (maximum zoom by a factor of 1.1), lighting adjustments (maximum adjustment by a factor of 0.2), and warping (maximum warping by a factor of 0.2). All transformations were applied with a probability of 0.75 except for horizontal flipping (probability = 0.5) using fastai.

The model was constructed using the U‐Net architecture (Ronneberger et al., 2015), which is known to be computationally efficient during training and highly accurate for segmentation in a diverse range of applications. Leveraging pretrained models using transfer learning, we used a dynamic U‐Net as implemented in fastai, building our model atop a resnet‐34 architecture pretrained for classification on the ImageNet data set (Deng et al., 2009). The pretrained resnet‐34 architecture was downloaded from the PyTorch model zoo (https://pytorch.org/docs/stable/torchvision/models.html).

We set aside 20% of our original images (n = 80) to validate our model. The pixels of these images were never presented to the model during training and thus model performance on these images can be viewed as a measure of the model’s applicability to novel data. We trained our model for 22 epochs (one epoch equals one pass through the training data), following the one‐cycle learning rate policy (Smith, 2018).

Segmentation model performance

We evaluated the performance of our model by comparing the predicted pixel labels from the U‐Net with the ground‐truth labels we generated for the 80 validation images. Across all 80 images in the validation set, our model achieved a 0.95 Sørensen–Dice coefficient on the predicted masks generated by the model (Fig. 3C). In this case, the Sørensen–Dice coefficient is equivalent to the proportion of shared pixel identities between the ground‐truth mask and the predicted mask. It is important to note that generating predictions for even a single herbarium image involves 65,536 predictions (i.e., 256 × 256 pixels). The performance of our model is therefore evaluated across ~5 million individual pixel predictions.

In order to understand how the model performed across the taxonomic diversity of our sample, we also measured the Sørensen–Dice coefficient for each of the 23 fern families in our validation data (Table 1). We found that the model performed roughly comparably across these groups, although for two families the model generated a Sørensen–Dice coefficient < 0.9. On inspection, the validation images representing these two families (Schizaeaceae and Athyriaceae) expose some key limitations of the model. First, the image representing the Schizaeaceae (catalog number 1054905, see Data Availability) contained very little leaf material and mostly thin stems. In the absence of leaves to outline, the model performance was evaluated entirely on its ability to partition thin stems from the backdrop. In general, our model may struggle with these types of features. The image representing Athyriaceae (catalog number 66902) was generated under less than ideal lighting conditions using older digitization technology and also shows a yellowed herbarium sheet. Model performance may be compromised under these conditions, although the 0.86 Sørensen–Dice coefficient for this image is still likely to meet the standard of quality for nearly all herbarium applications.

TABLE 1.

Model performance for individual fern families.

Family No. of validation images Sørensen–Dice coefficient
Gleicheniaceae 4 0.972
Lygodiaceae 4 0.966
Hymenophyllaceae 13 0.922
Equisetaceae 5 0.917
Ophioglossaceae 4 0.959
Marattiaceae 3 0.971
Psilotaceae 1 0.906
Osmundaceae 1 0.963
Schizaeaceae 2 0.882
Anemiaceae 2 0.952
Cyatheaceae 12 0.968
Polypodiaceae 5 0.948
Dryopteridaceae 4 0.948
Pteridaceae 6 0.963
Tectariaceae 1 0.916
Aspleniaceae 3 0.963
Lindsaeaceae 1 0.936
Blechnaceae 1 0.963
Thelypteridaceae 4 0.954
Athyriaceae 1 0.861
Salviniaceae 1 0.960
Dicksoniaceae 1 0.958
Marsileaceae 1 0.942

CONCLUSIONS

The deep learning model we present here can rapidly generate high‐quality masks of images of any herbarium sample across the morphological diversity of ferns. Although input images to the model may contain different color bars or palettes and other herbarium‐specific labels and features, the masks retain only the pixels of each image that belong to plant material, meaning the output masks can be combined into a single data set to allow computation across images from multiple herbaria. Our training data set was restricted to ferns; therefore, our model may not work as well for plant tissues absent from our data set (e.g., flowers and fruits). However, the workflow we present is general to herbarium images broadly and can be used by botanists specializing in other plant taxa to create their own set of masks with which to train similar deep learning models for segmentation. We hope that members of the botanical community with interests in specific taxa or tissues will create and share similar high‐resolution data sets and models.

There are ~18.9 million digitized herbarium images currently accessible through the Integrated Digitized Biocollections portal (iDigBio, https://www.idigbio.org/portal). This new scale of herbarium science requires modern tools and novel approaches to wield such massive data. Although automated segmentation is a large part of the preprocessing needed before a data set of herbarium images can be used for downstream deep learning applications, there are other concerns when using digital images of herbarium specimens not addressed in our workflow. For example, in any large data set, it is common to find multiple specimens in which the plant material is contained solely as fragments in envelopes and is therefore not visible. Deep learning may yet offer solutions to these challenges; however, workflows, training data, and models designed for preprocessing herbarium images for massive‐scale (millions of images) analyses are still needed to make deep learning tools accessible to the broader botanical community. It is critical to recognize the “human in the loop” component of machine learning and its applications in the biodiversity sciences. The extent to which botanists and biologists are willing to become familiar with these advanced computational tools will likely predict the utility of machine learning in botanical sciences as well as the novelty of the insights such tools may reveal.

Acknowledgments

Isabel Meyer, Duy Phan, and Sylvia Orli coordinated the transfer of images from the Smithsonian Digital Asset Management System to the Smithsonian Institution High‐Performance Cluster (SI/HPC). We thank Eric Schuettpelz and Larry Dorr at the National Museum of Natural History (Washington, D.C., USA) Department of Botany for their ongoing support. This work is supported by a Smithsonian Scholarly Studies Grant to R.B.D. and P.B.F. Image processing was performed on the SI/HPC (https://doi.org/10.25572/SIHPC) and the Brigham Young University Fulton Supercomputing Laboratory Cluster. We thank Emily Evans for her advice on image segmentation and Christine Frandsen for her assistance with Fig. 2.

White, A. E. , Dikow R. B., Baugh M., Jenkins A., and Frandsen P. B.. 2020. Generating segmentation masks of herbarium specimens and a data set for training segmentation models using deep learning. Applications in Plant Sciences 8(6): e11352.

Data Availability

Jupyter Notebook and Python script are available on GitHub: https://github.com/sidatasciencelab/fern_segmentation. Original images (https://doi.org/10.25573/data.9922148), masks (https://doi.org/10.25573/data.9922232), and metadata (https://doi.org/10.25573/data.11771004) are available on the Smithsonian Institution Figshare (smithsonian.figshare.com).

LITERATURE CITED

  1. Ba, J. , Mnih V., and Kavukcuoglu K.. 2014. Multiple object recognition with visual attention. arXiv 1412.7755 [Preprint]. Published 24 December 2014 [accessed 7 May 2020]. Available at: https://arxiv.org/abs/1412.7755.
  2. Bradski, G. , and Kaehler A.. 2008. Learning OpenCV: Computer vision with the OpenCV library. O’Reilly Media Inc., Sebastopol, California, USA. [Google Scholar]
  3. Burns, K. C. , Herold N., and Wallace B.. 2012. Evolutionary size changes in plants of the southwest Pacific. Global Ecology and Biogeography 21: 819–828. [Google Scholar]
  4. Choudhury, S. D. , Samal A., and Awada T.. 2019. Leveraging image analysis for high‐throughput plant phenotyping. Frontiers in Plant Science 10: 508. [DOI] [PMC free article] [PubMed] [Google Scholar]
  5. Ciresan, D. , Giusti A., Gambardella L. M., and Schmidhuber J.. 2012. Deep neural networks segment neuronal membranes in electron microscopy images. Advances in Neural Information Processing Systems 25: 2843–2851. [Google Scholar]
  6. Daru, B. H. , Park D. S., Primack R. B., Willis C. G., Barrington D. S., Whitfeld T. J. S., Seidler T. G., et al. 2018. Widespread sampling biases in herbaria revealed from large‐scale digitization. New Phytologist 217: 939–955. [DOI] [PubMed] [Google Scholar]
  7. Davis, C. C. , Willis C. G., Connolly B., Kelly C., and Ellison A. M.. 2015. Herbarium records are reliable sources of phenological change driven by climate and provide novel insights into species’ phenological cueing mechanisms. American Journal of Botany 102: 1599–1609. [DOI] [PubMed] [Google Scholar]
  8. Deng, J. , Dong W., Socher R., Li L.‐J., Li K., and Fei‐Fei L.. 2009. ImageNet: A large‐scale hierarchical image database, 248–255. 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, Florida, USA.
  9. Easlon, H. M. , and Bloom A. J.. 2014. Easy Leaf Area: Automated digital image analysis for rapid and accurate measurement of leaf area. Applications in Plant Sciences 2(7): 1400033. [DOI] [PMC free article] [PubMed] [Google Scholar]
  10. Garcia‐Garcia, A. , Orts‐Escolano S., Oprea S., Villena‐Martinez V., and Garcia‐Rodriguez J.. 2017. A review on deep learning techniques applied to semantic segmentation. arXiv 1704.06857 [Preprint]. Published 22 April 2017 [accessed 7 May 2020]. Available at: https://arxiv.org/abs/1704.06857.
  11. Gehan, M. A. , Fahlgren N., Abbasi A., Berry J. C., Callen S. T., Chavez L., Doust A. N., et al. 2017. PlantCV v2: Image analysis software for high‐throughput plant phenotyping. PeerJ 5: e4088. [DOI] [PMC free article] [PubMed] [Google Scholar]
  12. He, T. , Zhang Z., Zhang H., Zhang Z., Xie J., and Li M.. 2018. Bag of tricks for image classification with convolutional neural networks. arXiv 1812.01187 [Preprint]. Published 4 December 2018 [accessed 7 May 2020]. Available at: https://arxiv.org/abs/1812.01187.
  13. Heberling, J. M. , Prather L. A., and Tonsor S. J.. 2019. The changing uses of herbarium data in an era of global change: An overview using automated content analysis. BioScience 69(10): 812–822. [Google Scholar]
  14. Howard, J. , Gugger S., Bekman S., Ingham F., Monroe F., Shaw A., and Thomas R.. 2018. fastai. Website https://github.com/fastai/fastai [accessed 29 April 2020].
  15. Kavanagh, P. H. , Lehnebach C. A., Shea M. J., and Burns K. C.. 2011. Allometry of sexual size dimorphism in dioecious plants: Do plants obey Rensch’s rule? American Naturalist 178: 596–601. [DOI] [PubMed] [Google Scholar]
  16. LeCun, Y. , Bengio Y., and Hinton G.. 2015. Deep learning. Nature 521: 436–444. [DOI] [PubMed] [Google Scholar]
  17. Li, L. , Xu M., Wang X., Jiang L., and Liu H.. 2019. Attention based glaucoma detection: A large‐scale database and CNN model. arXiv 1903.10831 [Preprint]. Published 26 March 2019 [accessed 7 May 2020]. Available at: https://arxiv.org/abs/1903.10831.
  18. Lorieul, T. , Pearson K. D., Ellwood E. R., Goëau H., Molino J. F., Sweeney P. W., Yost J. M., et al. 2019. Toward a large‐scale and deep phenological stage annotation of herbarium specimens: Case studies from temperate, tropical, and equatorial floras. Applications in Plant Sciences 7: e01233. [DOI] [PMC free article] [PubMed] [Google Scholar]
  19. Meineke, E. K. , Davis C. C., and Davies T. J.. 2018. The unrealized potential of herbaria for global change biology. Ecological Monographs 88(4): 505–525. [Google Scholar]
  20. Meineke, E. K. , and Davies T. J.. 2019. Museum specimens provide novel insights into changing plant‐herbivore interactions. Philosophical Transactions of the Royal Society B 374: 20170393. [DOI] [PMC free article] [PubMed] [Google Scholar]
  21. Milioto, A. , Lottes P., and Stachniss C.. 2018. Real‐time semantic segmentation of crop and weed for precision agriculture robots leveraging background knowledge in CNNs, 2229–2235 In 2018 IEEE International Conference on Robotics and Automation in Brisbane, Queensland, Australia, May 2018. IEEE, Piscataway, New Jersey, USA. [Google Scholar]
  22. Moeskops, P. , Viergever M. A., Mendrik A. M., de Vries L. S., Benders M. J. N. L., and Isgum I.. 2016. Automatic segmentation of MR brain images with a convolutional neural network. IEEE Transactions on Medical Imaging 35(5): 1252–1261. [DOI] [PubMed] [Google Scholar]
  23. Otsu, N. 1979. A threshold selection method from gray‐level histograms. IEEE Transactions on Systems, Man, and Cybernetics 9(1): 62–66. [Google Scholar]
  24. Park, I. W. , and Mazer S. J.. 2018. Overlooked climate parameters best predict flowering onset: Assessing phenological models using the elastic net. Global Change Biology 24(12): 5972–5984. [DOI] [PubMed] [Google Scholar]
  25. Paszke, A. , Gross S., Chintala S., Chanan G., Yang E., DeVito Z., Lin Z., et al. 2017. Automatic differentiation in PyTorch. NIPS Autodiff Workshop. Website https://openreview.net/forum?id=BJJsrmfCZ [accessed 29 April 2020].
  26. Pearse, W. D. , Davis C. C., Inouye D. W., Primack R. B., and Davies T. J.. 2017. A statistical estimator for determining the limits of contemporary and historic phenology. Nature Ecology and Evolution 1: 1876–1882. [DOI] [PubMed] [Google Scholar]
  27. Ren, S. , He K., Girshick R., and Sun J.. 2015. Faster R‐CNN: Towards real‐time object detection with region proposal networks In Proceedings of the 28th International Conference on Neural Information Processing Systems, vol. 1, 91–99. [Google Scholar]
  28. Ronneberger, O. , Fischer P., and Brox T.. 2015. U‐Net: Convolutional networks for biomedical image segmentation. arXiv 1505.04597 [Preprint]. Published 18 May 2015 [accessed 7 May 2020]. Available at: https://arxiv.org/abs/1505.04597.
  29. Roth, H. R. , Farag A., Lu L., Turkbey E. B., and Summers R. M.. 2015. Deep convolutional networks for pancreas segmentation in CT imaging. arXiv 1504.03967 [Preprint]. Published 15 April 2015 [accessed 7 May 2020]. Available at: https://arxiv.org/abs/1504.03967.
  30. Rueden, C. Schindelin T. J., Hiner M. C., DeZonia B. E., Walter A. E., Arena E. T., and Eliceiri K. W.. 2017. ImageJ2: ImageJ for the next generation of scientific image data. BMC Bioinformatics 18: 529. [DOI] [PMC free article] [PubMed] [Google Scholar]
  31. Schneider, C. A. , Rasband W. S., and Eliceiri K. W.. 2012. NIH Image to ImageJ: 25 years of image analysis. Nature Methods 9(7): 671–675. [DOI] [PMC free article] [PubMed] [Google Scholar]
  32. Schuettpelz, E. , Frandsen P. B., Dikow R. B., Brown A., Orli S., Peters M., Metallo A., et al. 2017. Applications of deep convolutional neural networks to digitized natural history collections. Biodiversity Data Journal 5: e21139. [DOI] [PMC free article] [PubMed] [Google Scholar]
  33. Simonyan, K. , and Zisserman A.. 2014. Very deep convolutional networks for large‐scale image recognition. arXiv 1409.1556 [Preprint]. Published 4 September 2014 [accessed 7 May 2020]. Available at: https://arxiv.org/abs/1409.1556.
  34. Singh, A. , Ganapathysubramanian B., Singh A. K., and Sarkar S.. 2016. Machine learning for high‐throughput stress phenotyping in plants. Trends in Plant Science 21 2: 110–124. [DOI] [PubMed] [Google Scholar]
  35. Smith, L. N. 2018. A disciplined approach to neural network hyper‐parameters: Part 1 – learning rate, batch size, momentum, and weight decay. US Naval Research Laboratory Technical Report 5510‐026. arXiv 1803.09820 [Preprint]. Published 26 March 2018 [accessed 7 May 2020]. Available at: https://arxiv.org/abs/1803.09820.
  36. Thiers, B. M. 2018. The world’s herbaria 2017: A summary report based on data from Index Herbariorum. Website http://sweetgum.nybg.org/science/docs/The_Worlds_Herbaria_2017_5_Jan_2018.pdf [accessed 29 April 2020].
  37. Unger, J. , Merhof D., and Renner S.. 2016. Computer vision applied to herbarium specimens of German trees: Testing the future utility of the millions of herbarium specimen images for automated identification. BMC Evolutionary Biology 16: 248. [DOI] [PMC free article] [PubMed] [Google Scholar]
  38. van Rossum, G. , and Drake F. L.. 2009. Python 3 Reference Manual. CreateSpace, Scotts Valley, California, USA. [Google Scholar]
  39. Vasco, A. , Moran R. C., and Ambrose B. A.. 2013. The evolution, morphology, and development of fern leaves. Frontiers in Plant Science 4: 16. [DOI] [PMC free article] [PubMed] [Google Scholar]
  40. Willis, C. G. , Ellwood E. R., Primack R. B., Davis C. C., Pearson K. D., Gallinat A. S., Yost J. M., et al. 2017. Old plants, new tricks: Phenological research using herbarium specimens. Trends in Ecology and Evolution 32 7: 531–546. [DOI] [PubMed] [Google Scholar]
  41. Zhuang, Z. , Raj A. N. J., Jain A., Ruban N., Chaurasia S., Li N., Lakshmanan M., and Murugappan M.. 2019. Nipple segmentation and localization using modified U‐Net on breast ultrasound images. Journal of Medical Imaging and Health Informatics 9 9: 1827–1837. [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

Jupyter Notebook and Python script are available on GitHub: https://github.com/sidatasciencelab/fern_segmentation. Original images (https://doi.org/10.25573/data.9922148), masks (https://doi.org/10.25573/data.9922232), and metadata (https://doi.org/10.25573/data.11771004) are available on the Smithsonian Institution Figshare (smithsonian.figshare.com).


Articles from Applications in Plant Sciences are provided here courtesy of Wiley

RESOURCES