Skip to main content
Plant Phenomics logoLink to Plant Phenomics
. 2020 Nov 5;2020:5839856. doi: 10.34133/2020/5839856

“Macrobot”: An Automated Segmentation-Based System for Powdery Mildew Disease Quantification

Stefanie Lück 1, Marc Strickert 2, Maximilian Lorbeer 3, Friedrich Melchert 4, Andreas Backhaus 4, David Kilias 4, Udo Seiffert 4, Dimitar Douchkov 1,
PMCID: PMC7706317  PMID: 33313559

Abstract

Managing plant diseases is increasingly difficult due to reasons such as intensifying the field production, climatic change-driven expansion of pests, redraw and loss of effectiveness of pesticides, rapid breakdown of the disease resistance in the field, and other factors. The substantial progress in genomics of both plants and pathogens, achieved in the last decades, has the potential to counteract this negative trend, however, only when the genomic data is supported by relevant phenotypic data that allows linking the genomic information to specific traits. We have developed a set of methods and equipment and combined them into a “Macrophenomics facility.” The pipeline has been optimized for the quantification of powdery mildew infection symptoms on wheat and barley, but it can be adapted to other diseases and host plants. The Macrophenomics pipeline scores the visible powdery mildew disease symptoms, typically 5-7 days after inoculation (dai), in a highly automated manner. The system can precisely and reproducibly quantify the percentage of the infected leaf area with a theoretical throughput of up to 10000 individual samples per day, making it appropriate for phenotyping of large germplasm collections and crossing populations.

1. Introduction

Cereals, which include wheat, barley, rice, maize, rye, oats, sorghum, and millet, have been the primary component of humans' diet delivering more than 50% of the world's daily caloric intake [1]. Like any other plant, these species are under constant attack by a vast number of pathogens. However, because the impact of cereal diseases is proportional to the importance of these crops for human nutrition, they are of exceptional interest to plant pathologists and breeders.

Precise and sensitive phenotyping is one of the critical requirements for modern breeding and functional genomics studies. Many of the desired traits are polygenic by nature, and their manifestation depends on the cumulative effect of several factors with a small to moderate effect. The quantitative disease resistance of the plants against pathogens is a typical example of a complex polygenic trait. Although this type of resistance is usually less efficient than the strong R-gene-based resistance, it is nevertheless a desired trait because of its durability on the field, and in high contrast to the R-gene resistance, it is effective against all races of a particular pathogen and even against different pathogen species. However, studying the underlying mechanisms of the quantitative resistance is seriously challenged by the complexity of this phenomenon [2, 3]. The accessibility of the genomic information for several host and pathogen species greatly facilitates these studies but, on the other hand, introduced an enormous amount of data that needs to be tested and functionally validated. Thus, the ability of high throughput becomes an essential requirement for the new systematic phenotyping, and the term “phenomics” was coined to describe this approach.

The natural disease resistance is, besides the high yield and abiotic stress resistance, one of the most desired crop traits since the beginning of the agriculture. The breeders invested significant efforts in improving these traits, and as a result, the modern crop cultivars are usually outperforming their wild progenitors in nearly all aspects. However, unlike other factors that may influence plant performance, the pathogens actively develop and modify strategies to evade the host defense mechanisms in a process sometimes called “evolutionary arms race”.

Powdery mildew (PM) is a disease caused by a diverse group of obligate biotrophic fungi that lead to extensive damage to various crop plants, including cereals. Blumeria graminis is the causative agent of the powdery mildew disease of wheat and barley [4]. Like most of the obligate biotrophs, B. graminis shows extreme host specificity. The so-called formae speciales (f.sp.) have specialized virulence for particular plant species, e.g., for barley (B. graminis f.sp. hordei) or wheat (B. graminis f.sp. tritici).

The asexual life cycle of B. graminis, from the beginning of the infection to the production of new spores, completes within a week. The haploid asexual fungal spores, called conidia, start germination within a few hours after contact with a plant leaf. The appressorial germ tube penetrates the cell wall of the leaf epidermal cells directly and grows into the living plant cell forming a feeding structure called a haustorium. The establishment of biotrophy occurs within the first 24 hours after leaf spore inoculation. In the following days, epiphytically growing hyphae develop many secondary haustoria in neighboring epidermal cells next to the initial infection site. After three days, the fungal colony is macroscopically visible. In the following days, abundant spores are formed by the mycelium, which completes the life cycle [5]. In controlled infection assays with defined spore titters, the severity of the infection and the size of the infected area are commonly the scoring parameters in disease rating to estimate host susceptibility [6].

Cultivated barley (Hordeum vulgare spp. vulgare), a member of the Triticeae tribe of grasses, is among the most favored crops worldwide. Besides, barley is a famous genetic model for the very closely related but more complex wheat genome.

With the significant progress made on the sequencing of several cereal genomes, Genome-Wide Association Studies (GWAS) to identify resistance traits became possible. However, a bottleneck for successful genotype-phenotype associations is the high-throughput monitoring of disease symptom development as a measure of host plant susceptibility. Disease resistance traits range from partial, or quantitative, to complete, or qualitative. It has been shown in many cases that quantitative disease resistance is more durable on the field and, therefore, of high potential value to the breeders [7]. However, the quantitative resistance is usually a polygenic trait, which is based on the joined effect of many genes, where each of them contributes quantitatively to the level of plant defense [8]. The identification of genes with small to moderate resistance effects requires exact and reproducible quantification of infection as a prerequisite for genetic fine mapping and gene isolation.

The choice of high-throughput phenotyping technologies for disease resistance has rapidly increased over the last years. There are methods based on measuring the enzymatic activity of the infected tissue [9], on chlorophyll fluorescence [10], or on quantitative PCR of fungal genes [11], but more commonly optical sensors and computer vision approaches are used [12]. Hyperspectral imaging is using the information about the reflectance of the tissues in a wide range of wavelengths and may visualize the disease symptoms in relatively early stages [13, 14]. Multispectral imaging is done with only a few but usually highly informative wavelengths, thus significantly reducing the cost of equipment and the amount of raw data. However, the most common type of optical sensors is using the visible and near-visible spectrum. These sensors are either with integral wide-band filter matrices for limiting the sensitivity on the pixel level to specific wavelengths (e.g., RGB cameras) or without wavelength discrimination (grayscale cameras) but often with external filters and/or illumination sources with a discrete wavelength band.

Although the phenomics platforms are typically built on highly customized hardware, some implementations are using standard hardware components from the consumer market. This approach allows the building of very cost-efficient and versatile systems but with certain limitations. An example of such a system represents the PhenoBox [15], which is based on a consumer mirror-reflex camera for phenotyping in controlled conditions. The system allows phenotyping of biotic and abiotic stress of small plants and plant organs at a very moderate cost. However, the inbuilt filters of the used camera type allow imaging only in the visible range of light.

Designs based on multiple single-board computers (e.g. Raspberry Pi and Arduino) provide another exciting low-cost alternative to the large robotics platforms. An example of such a system is the low-cost SeedGerm platform [16], which is using multiple Raspberry Pi computers and implements a single-axis camera movement for seed imaging and germination phenotyping. Nevertheless, typically the low-cost devices are usually missing robotics components for sample handling that renders them less appropriate for high-throughput and 24/7 applications.

The development of image analysis methods is very dynamic and contributes to increasing the expectations for more complex phenotypes. Currently, two primary approaches are commonly used to analyze image data. The first one is using custom handcrafted features, and the other one is based on Artificial Neural Networks (ANN). The first approach typically requires a significant manual input in the form of selecting and implementing informative features, but efficient models can be built with a relatively low amount of training data, which makes this approach attractive in some instances [17, 18].

The current state of the art machine learning methods is based on ANN and most recently deep learning (DL) architectures [1922]. However, early DL architectures are less appropriate for semantic segmentation (pixel-based classification) because of the missing spatial information, which is an essential feature for conventional image recognition networks such as convolutional neuronal networks [23]. Therefore, edge detection, clustering, or thresholding-based segmentation solutions are still frequently used for pixel classification [24]. Another drawback of the DL is the typically very high demand for annotated training samples and high hardware requirements. New Convolutional Neural Network (CNN) architectures like Mask Regional-CNN (Mask R-CNN) [25] or U-Nets CNN [26] promise to solve many of the limitations of the DL and are already used for semantic segmentation of plant diseases [27, 28].

This work is aimed at establishing a high-throughput, automated phenotyping platform for precise and reproducible quantification of leaf disease of cereals, with a focus on powdery mildews and rusts.

2. Materials and Methods

2.1. Experimental Design

The Macrophenomics pipeline consists of hardware and software components. A specialized robotic system (Macrobot) implements the image acquisition part of the Macrophenomics pipeline. The Macrobot autonomously acquires images of detached leaf segments mounted on standard size microtiter plates (MTPs) (Figure 1).

Figure 1.

Figure 1

Overview of the phenotyping pipeline. The plants are grown in 24-well trays in a greenhouse. At the appropriate stage, leaf fragments are harvested and mounted on standard 4-well microtiter plates, filled with 1% water agar for keeping the humidity, and inoculated by air-blowing of powdery mildew spores in an inoculation tower. After incubation of 5-7 days, the disease symptoms become visible. The plates with the infected leaves are loaded into the Macrobot system for automated imaging. The acquired images are transferred to the image analysis server for quantification of the disease symptoms.

Typically, the wheat and barley plants are grown in 24-well trays in a greenhouse. The samples are taken at the 2-leaf stage from the middle part of the second leaf. The leaf fragments are mounted on standard 4-well MTPs with 1% water agar (Phyto agar, Duchefa, Haarlem, the Netherlands) supplemented with 20 mg L−1 benzimidazole as a leaf senescence inhibitor. For achieving regular inoculation of all leaves, the plates without lids are placed in a rotating table inside an inoculation tower and are inoculated by blowing in conidiospores from sporulating material. Inoculated plates are incubated in environmentally controlled plant growth chambers (20°C, 60% RH constant; 16 h light, 15 μE m−2 s−1) for 6 days until the disease symptoms are visible. The infected plates are loaded into the Macrobot system for automated imaging. The acquired images are transferred to the image analysis server for quantification of the disease symptoms.

2.2. Hardware

In the original version, the Macrobot employs a 14-bit monochrome camera (Thorlabs 8050M-GE-TE) at a resolution of 3296 × 2472px. A high-end lens (CoastalOpt UV-VIS-IR 60 mm 1 : 4 Apo Macro) with apochromatic correction in the range from 310 to 1100 nm wavelength ensures that images using different illumination setups are precisely registered and focused. The illumination is realized using small-bandwidth isotropic LED light sources (Metaphase Exolight-ISO-14-XXX-U) with 365 nm (UV), 470 nm (blue), 530 nm (green), and 625 nm (red) peak wavelengths.

The robotic part of the Macrobot is built of off-the-shelf OEM components developed for laboratory use that facilitate the maintenance and allow quick replacement of defective components. Specifically, a PlateCrane EX Microplate Handler (Hudson Robotics, Inc., NJ, USA) in conjunction with an IGUS precision linear stage (igus GmbH, Germany) is used. The plate crane and linear stage can be accurately positioned with high repetition accuracy using the vendors' software API. The positions of the plate crane and linear stage for the different stages of plate handling are determined at the initial system setup and have to be manually readjusted if the system was moved or rearranged in some form. Positional information is stored in parameter files and reloaded in the Fraunhofer IFF custom implemented system software. This also applies to the imaging settings of the Thorlabs camera. The imaging process is controlled by the system software using its custom scripting engine, which allows a controlled complex sequence of steps to unfold in order to ensure a repeatable and high-throughput measurement on the supplied sample set. Imaging procedure parameters can be adjusted by the user and are stored as profiles to ensure repeatable imaging.

For each plate, monochrome images in all illumination wavelengths are acquired separately and stored in 16-bit TIFF image files. An RGB image is generated by combining the images of the red, green, and blue LED channels (Supplemental Figure S1). The UV channel is used to facilitate the extraction of the region of interest (ROI), where the leaves are located. Video sequences showing the Macrobot in action can be seen in [29].

An improved version of Macrobot was introduced on a later stage and designated as Macrobot 2.0 (Figure 2). The illumination system was upgraded by doubling the LED units allowing bilateral illumination of the objects. A background illumination system based on electroluminescence foil was mounted on the MTP carrier to simplify the separation between the foreground and background, thus improving the leaf segmentation. The image acquisition and hardware controlling software was upgraded to a 64-bit version for optimal system memory utilization. The entire technical layout was improved with respect to the gained experience with the first version of the Macrobot. Since the image acquisition components remain unchanged, data generated by Macrobot 2.0 is fully comparable to data acquired by Macrobot 1.0, as far as a comparable hardware setup is used (e.g., one-sided illumination). The data presented in this article was acquired by the original Macrobot hardware configuration.

Figure 2.

Figure 2

Macrobot 2.0 with improved technical design, bilateral illumination, and background light: (a) outside view and (b) inside view of the photo box.

2.3. Software

The image analysis software was implemented in Python 3.8 under Microsoft Windows 10 with extensive use of the NumPy (v. 1.12.1) [30], opencv-python (v. 2.4.13), scikit-learn (v. 0.17.1) [31], and scikit-image (v. 0.13.0) [30] open-source libraries. The source code is available at [32].

2.4. Model Evaluation

Each model was validated by calculating the accuracy, recall, and precision of the model to test the prediction performance for each class. The overall accuracy is calculated by the number of correctly predicted observations divided by the total number of observations:

Accuracy=TP+TNTP+FP+FN+TN. (1)

Precision is a measure of the false-positive rate. It can be calculated by dividing the true positive observations by the total predicted positive observations:

Precision=TPTP+FP. (2)

Recall measures the sensitivity of the predicted positive observations:

Recall=TPTP+FN. (3)

2.5. Wheat Genotype Collection

A collection of 188 elite lines and 202 genetic resource lines was obtained from the federal ex situ collection of the Gene Bank of IPK Gatersleben. The complete list of the genotypes is given in Supplemental Table S6.

2.6. Best Linear Unbiased Estimations (BLUEs) of Genotype Response against Fungal Infections

The phenotypic results of the tested genotype collection are represented as BLUEs obtained by fitting the data to the following linear mixed model:

y=1nμ+XGg+ZRr+e, (4)

where y is the vector containing the n original raw fungal infection data points, 1n indicates an n-size vector of only 1 s, μ stands for the fixed intercept term, g is a vector of the fixed genotypic effects of the material being tested, and r represents a vector that contains all random factors beyond residual variation: experiment, treatment, the interaction between experiments and genotypes, and the interaction between genotypes and treatments, while e indicates a vector of random residual variation. XG and ZR are design matrices that assign g and r to the corresponding values contained within y. Mixed model equations were computed using the lme4 package [33] implemented in R Environment ver. 3.4.0 [34].

2.7. Plant and Fungal Material

Wheat and barley plants from different cultivars and landraces were grown in 24-pot trays (31 × 53cm) in a greenhouse at 20°C constant and 16 h light period in a soil substrate. The first or the second leaves were harvested at 7 days and 13-14 days, respectively, after sowing. The leaf segments were mounted on 20 mg L−1 benzimidazole-supplemented and 1% water agar plates and inoculated with the corresponding pathogen at approximately 10 spores/mm2. As pathogens, the Swiss wheat powdery mildew field isolate FAL 92315 and the Swiss barley powdery mildew field isolate CH4.8 were used, respectively. The image acquisition was performed seven days after inoculation (dai).

2.7.1. Wheat Genotype Collection

A collection of 187 elite lines and 201 plant genetic resource lines was obtained from the federal ex situ collection of the Gene Bank of IPK Gatersleben. The complete list of the genotypes is given in Supplemental Table S6.

2.8. Quantitative PCR

Quantitative real-time PCR was performed in a volume of 5 mL QuantiTect Probe PCR Kit (Qiagen GmbH, Hilden, Germany) and an ABI 7900HT fast real-time PCR system (Thermo Fisher Scientific Inc., Waltham, MA, USA). Forty cycles (15 sec, 94°C; 30 sec, 56°C; 30 sec, 72°C, preceded by standard denaturation steps at 94°C for 2 min) were conducted. Data were analyzed by the standard curve method using the SDS 2.2.1 software (Thermo Fisher Scientific Inc., Waltham, MA, USA). A standard curve dilution series was included for each gene, as fivefold dilutions and three technical replicates per DNA sample. The detected quantity of the fungal gene GTFI (beta-1,3-glucanosyltransferase, GenBank: EU646133.1) was normalized to the quantity of the barley UBC gene (ubiquitin-conjugating enzyme, GenBank: AY220735.1) and used as a proxy for fungal biomass.

The used primers and probes are as follows: for the powdery mildew GTFI gene—BgGTF1_F (5′TTGGCCAAACAACTCAACTC3′), BgGTF1_R (AGCAGACCAAGACACACCAG), and BgGTF1_PR (fluorescent TaqMan probe, FAM-5′CTCCCAGCAACACTCCAGCT3′-BHQ1), and for the barley UBC gene—HvUBC_F (5′ACTCCGAAGCAGCCAGAATG3′), HvUBC_R (5′GATCAAGCACAGGGACACAAC3′), and HvUBC_PR (fluorescent TaqMan probe Yakima Yellow-5′GAGAACAAGCGCGAGTACAACCGCAAGGTG3′-BHQ1).

3. Results

3.1. Frame and Leaf Segmentation

To define the area where the leaf segments are located on the plates, the C-shaped white frames that hold the leaves were segmented and extracted. Optimal results were achieved by applying Otsu's thresholding [35] on the UV channel, followed by dilation with an 8 × 8 kernel to obtain a binary image (Figure 3). The Moore-Neighbor [36] tracing was used to extract the contours of the binary image and filter the frames by size and position.

Figure 3.

Figure 3

Frame and leaf image segmentation processing chain. In the first step, the white leaf-holding frames are used to define the regions of interest (ROIs) that contain the leaf fragments. Next, the plate image is split on four ROIs, each one containing a row of leaf samples that typically belong to a single genotype but to a different individual plant (e.g., each leaf fragment in a row is a technical replicate). In parallel, the back-illuminated image of the plate is used to separate the leaves of each other. The final leaf segmentation is done by combining the row and leaf ROIs.

Each leaf segment was extracted to a separate region of interest (ROI). The best segmentation results were obtained by Otsu's binarization method on the backlight image, followed by the Moore-Neighbor contour finding algorithm and object size selection. Otsu's method also gets along with the particular challenge of interrupted leaf contours caused by necrosis or fungal infections.

3.2. Machine Learning Approach

The application of machine learning approaches gives the advantage of using a data-driven analysis rather than hypothesis-driven statistics. In this way, complex statistical modeling assumptions can be reduced, offering possibly important data features from which machine learning tools can derive desired classification outcomes in the manner of teaching. Therefore, several machine learning methods were implemented and evaluated for their accuracy and performance in the quantification of the PM disease symptoms.

3.2.1. Training Data

Training data was collected by manual labeling of background, infected, and leaf necrosis areas. The single labeled pixels were extracted and assigned to these three classes. For avoiding a class imbalance, the number of training samples per class was adjusted to the lowest number of pixels per class, which was 5000. The dataset was split: 70% for training and 30% for validation.

3.2.2. Feature Extraction and Classification

We have compared three conventional classifiers: C-Support Vector Classification [37], Linear Support Vector Classification [38], and random forest [39]. We found the random forest to be performing significantly better than the Linear Support Vector Classification and slightly better compared to the C-Support Vector Classification (Figure 4(a), Supplemental Table S2). The training time with the random forest classifier was about ten times faster than that with the C-Support Vector Classification. Therefore, we end up using the random forest classifier for further experiments.

Figure 4.

Figure 4

(a) Evaluation of different classifiers on HSV_H_channel (5000 pixels/class, n = 10, errorbars = SD). (b) Evaluation of the random forest classifier with different numbers of trees on HSV_H_channel (5000 pixels per class, n = 10).

To find the optimal number of trees for the random forest classifier, we tested six different values ranging from 10 to 200 trees, which lead to an optimum of 50 trees (Figure 4(b), Supplemental Table S3).

A random forest classifier has been trained by using RGB, Lab, and HSV as multiple- and single-color channels (Figure 5(a), Supplemental Table S4).

Figure 5.

Figure 5

(a) Evaluation of different color pixel classification methods (5000 pixels/class, n = 10, errorbars = SD). Random forest classifier (nr_trees = 50). (b) Evaluation of texture features (5000 pixels/class, n = 10, errorbars = SD). Random forest classifier (nr_trees = 50).

Texture spatial features such as local binary pattern [40], Haralick [41], and Parameter-Free Threshold Adjacency Statistics (PFTAS) [42] were also tested for improving the performance of the classifier (Figure 5(b), Supplemental Table S5).

Four models reached an overall accuracy above 0.80: the blue channel of the RGB color space, the hue channel of the HSV color space, the a channel of the Lab color space, and the Haralick texture features. Those models were tested further in the validation experiment.

3.3. Segmentation Approach

In addition to the machine learning approach, we have tested several segmentation methods: edge detection, superpixel segmentation, watershed transformation, region growing methods, thresholding, and minimum and maximum RGB (data not shown). The most efficient segmentation was achieved by the relatively simple method of minimum RGB (minRGB) (Figure 6). The algorithm takes the single values for each RGB channel, determines the minimum number of each channel, and stores the value. The other two channels are set to the value 0. This simple filter allowed reliable differentiation of the disease symptoms from the background by simultaneous reduction of the analysis artifacts and hardware workload.

Figure 6.

Figure 6

Minimal RGB value approach for segmentation.

3.4. Validation Experiment

The Macrophenomics module is aimed at providing a precise and reproducible evaluation of the experimental results and at the same time at releasing the human personnel from a routine and laborious task. To estimate the performance of the different approaches and computer models on independent material, we have carried out a validation experiment, where six domain experts were asked to do a manual disease rating. Combining the scores given by all experts formed a robust mean value, which was used to validate the computer prediction results. The validation set included a partially very difficult to score material with a lot of leaf senescence and necrosis.

In parallel to the visual methods, two other types of measurements were included for comparison: quantification of the total fungal biomass using quantitative real-time PCR (qPCR) of fungal DNA and inoculum density as the number of applied fungal spores per mm2 of leaf surface (Figure 7(a), Supplemental Table S1).

Figure 7.

Figure 7

(a) Plots of the infection area determined automatically (blue triangles) and mean manual values (“Mean man.,” green rectangles), together with the fungal biomass measured by qPCR (normalized relative transcript levels multiplied by 100 for better visibility, purple crosses) and inoculation density (spores per mm2, red rectangle, sorted ascending). On the x-axis are the ordered samples, and on the y-axis are the infection area (% of the leaf surface), relative fungal biomass (relative units), and inoculation density (spores/mm2). (b) The minimal and maximal visual infection scores (black bars) and the means (red dots) estimated by the domain experts. The graphs show the discrepancy of visual scoring of the involved persons. On the x-axis are the ordered samples sorted by the mean, and on the y-axis is the infection area (% of the leaf surface).

3.4.1. Performance of the Machine Learning Approach

All machine learning models with accuracy above 0.8 (Figures 4 and 5) plus the minRGB segmentation algorithm were tested on the validation experiment data by comparing the mean visual scores and the predictions of the corresponding models (Figure 8).

Figure 8.

Figure 8

Scatter plots and Pearson's coefficients of correlations (r) for the different machine learning models plus the minRGB-based segmentation (y-axis) versus the mean visual scores given by six experts (x-axis). PM infected detached barley leaves, 6 dai. The number of samples (n = 108).

The best correlation to the mean visual data (0.885) was achieved by the minRGB method, confirming once more its efficiency for image segmentation of the particular material.

3.4.2. Performance of the Segmentation Approach

Although the leaf material of the validation experiment was often covered by large necrotic and/or chlorotic areas, which may complicate the disease recognition even for domain experts, the minRGB-based prediction was very accurate (Figure 9).

Figure 9.

Figure 9

An example of a minRGB filter-based detection of the disease area on PM-infected detached barley leaves 6 dai. (a–c) RGB composite image, minRGB filter results, and final prediction after thresholding.

The minRGB-based algorithm was also tested in a larger experiment with wheat, showing an even higher level of accuracy (Figure 10).

Figure 10.

Figure 10

Scatter plot and Pearson's coefficients of correlation (r) between the visual score given by an expert (y-axis) and minRGB-based automated scores for the PM-infected area (x-axis), 6 days after infection. Number of samples (n = 660) and number of testers (p = 1). The manual scores were given in 5% steps.

The better results for the wheat material might be explained with the lower frequency of appearance of problematic artifacts such as necrosis and senescence in this particular material.

The run time per sample and per dataset was reduced up to 10-fold by using the minRGB approach in comparison to the per-pixel classification methods. With the particular hardware configuration, the image analysis time was up to 3-fold shorter than the time required for image acquisition, thus allowing the implementation of image analysis in real time.

3.5. Application of the Macrophenomics Platform for Evaluation of Quantitative Disease Resistance of Genetically Diverse Plant Material

The disease phenotype may strongly depend on the genotype of the host plant. Therefore, we have tested a collection of 388 genetically diverse wheat genotypes for quantitative disease resistance against wheat powdery mildew. The collection consists of 187 elite breeding lines, released post-Second World War, and 201 plant genetic resources, mostly historically collected landraces. The plants were grown to the two-leaf stage (13 days), and the second leaf was cut and inoculated with the wheat powdery mildew isolate FAL 92315. The disease symptoms were scored after 6 days of incubation at 20°C, 60% RH, and 16 h light period. Each genotype was tested in two biological repetitions, with six individual plants per genotype.

The result showed a very different distribution of the levels of quantitative disease resistance within the two groups, with a significant enrichment of resistance genotypes in the elite lines (Figure 11).

Figure 11.

Figure 11

Distribution of the BLUEs for powdery mildew disease resistance in 5250 individual plants from two biological replicates. The elite line pool consists of 187 genotypes; the plant genetic resources (landraces) consist of 201 genotypes. The higher BLUEs indicate higher susceptibility to the pathogen. The elite lines show significant enrichment of disease-resistant genotypes.

4. Discussion

A massive body of data on powdery mildew disease resistance of different plant genotypes was collected over the years by the breeders and researchers. Although a precious resource, the majority of this data is hardly reproducible because it is collected mostly in uncontrolled field conditions, with an unknown mix of pathogen isolates, and visually scored by many different persons. To overcome these problems and to establish a platform for precise phenotyping under controlled conditions, we have developed a high-throughput platform for phenotyping of cereal leaf diseases. The system is based on the well-established detached leaf assay [4345], which allows a very high level of controlling the environment and pathogen pressure. The method is also very well adapted for phenotyping via optical sensors since the leaves can be mounted in special containers for acquiring images in fully controlled conditions.

In our system, we have selected a monochrome CCD sensor to avoid some of the inbuilt problems of the RGB cameras (e.g., pixel value interpolation and lowered quantum efficiency). Instead of using filters for specific wavelengths, we decided to use narrow-bandwidth isotropic LED light sources, thus avoiding the use of motorized filter magazines and losing quantum efficiency. The nature of the samples (nonmoving fixed objects) allows the acquisition of several images per object and the combination of the data without complicated merging methods. The leaf samples are fixed in standardized containers (microtiter plates), which greatly simplify the hardware design allowing the use of commercially available components such as the plate crane. The white plastic frames that keep the leaves fixed in the plates are at the same time used to define the area of interests, where the leaves are located.

The central hardware part of the system is the Macrobot. It is equipped with custom imaging system software developed by Fraunhofer IFF (Magdeburg, Germany). Several software modules control all actors and sensors in the system providing services to a service manager. The flow control for the imaging process is achieved by script programming, which enables a change in the imaging process without reimplementing the different software modules and makes extensions to the system secure and efficient. System modules providing a graphical user interface are organized in a reconfigurable user interface, which can be arranged to the needs of the system user without reimplementation. The imaging system generates a structured dataset for the subsequent image analysis.

We have tested several machine learning and segmentation approaches to find the most efficient algorithm for disease quantification. The most informative features for the machine learning approach were the H, B, and a channels of the HSV, RGB, and Lab color spaces, respectively. Among the tested texture features, Haralick was the by far most informative. A combined pixel classification based on color and texture features was tested as well but without significant improvement compared to the single features. Of the three different classifiers that we have evaluated, the random forest (RF) performed slightly better than the Support Vector Classifier (SCV) and much better than the Linear SVC. We have tested also RF with a different number of trees, and we found the number of 50 trees to be optimal.

Astonishingly, among all tested segmentation approaches, the most accurate and efficient technique was the simple method of minimum RGB (minRGB). This filter was able to detect the infected leaf area reliably and to reduce the signal from disease-unrelated necrotic brown spots, which were of a particular problem in nearly all other approaches. Besides, the hardware workload and the calculation time for computing the minRGB filter were significantly lower than those of any other method. Finally, the minRGB was the segmentation method of choice, which was implemented into the image analysis pipeline.

A take-home message from this result can be that the use of sophisticated image analysis methods should not become a goal per se. In several cases, more straightforward methods may provide comparable results at a much lower cost.

We have validated the prediction results by three other direct and indirect quantification methods: a visual scoring as the mean value of the scores of six different domain experts, quantitative PCR (qPCR), and inoculum density as the number of spores per square millimeter of the leaf surface.

Although the genomic qPCR provides a nearly direct estimation of the total fungal biomass, it is a complicated method, which is influenced by many factors, such as genomic DNA isolation and quality, primer design, PCR efficiency, and detection sensitivity. Also, the measured quantity depends on both visible (on the leaf surface) and invisible (too small or internal) fungal structures and is therefore not necessarily in perfect correlation with the visible disease symptoms. The inoculum density is instead an indirect parameter, which gives the infection pressure and the potential for the formation of fungal colonies. However, the formation of the final fungal biomass depends on several other biotic and abiotic factors, such as spore fitness and aggressivity, plant response, and support of the fungal growth and temperature and humidity. The mean scoring value of several persons provides a very robust parameter, and therefore, it was the method of choice for calibration of the automatic prediction.

We have applied the Macrobot platform for the evaluation of an extensive, diverse collection of genotypes. The collection contains modern breeding material (elite lines) and a historical collection of landraces and semiwild genotypes. Widespread opinion among the nonexperts and some experts is that the wild species and landraces are typically more resistant to biotic and abiotic stress than the modern high-yielding varieties. Interestingly, our observation demonstrates that the elite lines showed a substantial enrichment of resistant genotypes compared to the landraces. This observation reflects the long-term efforts of the breeders for introducing disease resistance alleles in the elite material. These results prove the capability of the Macrobot system to deliver high-quality data for very diverse plant material. The obtained phenotypic data can be used directly for breeding purposes or GWAS approaches.

In this work, we demonstrate that our Macrophenomics platform can provide reliable and reproducible data in an excellent correlation with the classical scoring methods, and it can even outperform the scores of individual experts by the accuracy of infection area estimation. The platform is also fully open for adaptation to diseases other than powdery mildew leaf diseases such as different spot, blight, and rust diseases caused by several fungal, viral, and bacterial pathogens, such as yellow and brown rusts (Puccinia sp.), septoria leaf blotch (Zymoseptoria tritici), spot blotch (Bipolaris sorokiniana), bacterial leaf blight (Pseudomonas syringae), bacterial leaf streak and black chaff (Xanthomonas translucens), and barley yellow dwarf virus. However, an important limitation is that the tested objects must fit into a standard MTP container (app. 12 × 8 × 1cm), which includes samples like detached leaves, seeds, stem and root fragments, cereal spikes, and small whole plants.

A FAIR principle [46] compliant data management is currently under development. The pipeline follows the basic schema of the BRIDGE Visual Analytics Web Tool for Barley Genebank Genomics [47]. It will allow visualization and export of the phenotypic and genotypic data, online GWAS, and several other features of interest to the scientist and plant breeders.

Acknowledgments

We would like to express our sincere appreciation to Patrick Schweizer, who passed away before the submission of this manuscript and, according to the publisher's policy, is therefore not eligible to be a coauthor. Nevertheless, we have to acknowledge the major contribution of Patrick Schweizer to developing the idea and design of the Macrobot and the design and the interpretation of the experiments in this manuscript. We would like to acknowledge also the following members of the former Pathogen-Stress Genomics group at IPK Gatersleben: G. Brantin, Dr. W. Chen, Dr. D. Nowara, and Dr. J. Rajaraman for their contribution to the visual disease scoring data. Furthermore, thanks are due to Dr. D. Nowara for providing the primers and probes for the qPCR experiment, to Dr. Albert W. Schulthess for the helpful discussions on the statistical methods, and Dr. Armin Djamei for the critical reading of the manuscript and the valuable suggestions. This work was performed in the frames of the German Plant Phenotyping Network (DPPN) (FKZ 031A053) and GeneBank 2.0 project (FKZ 031B0184) funded by the German Federal Ministry of Education and Research (BMBF).

Data Availability

Image data used for validation of the Macrobot algorithm is available at [48]. The image analysis software is available at [32].

Conflicts of Interest

The authors declare that there is no conflict of interest regarding the publication of this article.

Authors' Contributions

SL designed and programmed the image analysis software and performed the validation experiments. MS contributed to writing the paper and to testing the machine learning approaches. ML performed the wheat infection experiment. FM, AB, DK, and US designed and developed the Macrobot hardware and controlling software. DD contributed to designing of the image analysis software and hardware, performed the experiments, and wrote the manuscript.

Supplementary Materials

Supplementary Materials

Supplemental Figure S1: schematic drawing of the image acquisition hardware of the Macrophenomics module Macrobot 2.0 (top and side views). Supplemental Table S1: validation experiment, all data (Excel sheet). Supplemental Table S2: values in Figure 4(a)—evaluation of different classifiers on HSV_H_channel (5000 pixels per class, n = 10). Supplemental Table S3: values in Figure 4(b)—evaluation of the random forest classifier with a different number of trees on HSV_H_channel (5000 pixels per class, n = 10). Supplemental Table S4: values in Figure 5(a)—evaluation of different color pixel classification methods (n = 10). Supplemental Table S5: values in Figure 5(b)—evaluation of texture features (n = 10). Supplemental Table S6: list of wheat genotypes (Excel sheet).

References

  • 1.Awika J. M. ACS Symposium Series. American Chemical Society; 2011. Advances in Cereal Science: Implications to Food Processing and Health Promotion. [DOI] [Google Scholar]
  • 2.Corwin J. A., Kliebenstein D. J. Quantitative resistance: more than just perception of a pathogen. Plant Cell. 2017;29(4):655–665. doi: 10.1105/tpc.16.00915. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Jones J. D. G., Dangl J. L. The plant immune system. Nature. 2006;444(7117):323–329. doi: 10.1038/nature05286. [DOI] [PubMed] [Google Scholar]
  • 4.Bockus W. W., Bowden R. L., Hunger R. M., Morrill W. L., Murray T. D., Smiley R. W. Compendium of Wheat Diseases and Pests. Third. Amer Phytopathological Society; 2010. [Google Scholar]
  • 5.Jankovics T., Komáromi J., Fábián A., Jäger K., Vida G., Kiss L. New insights into the life cycle of the wheat powdery mildew: direct observation of ascosporic infection in Blumeria graminis f. sp tritici. Phytopathology. 2015;105(6):797–804. doi: 10.1094/PHYTO-10-14-0268-R. [DOI] [PubMed] [Google Scholar]
  • 6.Nicot P. C., et al. The Powdery Mildews. In: Bélanger R. R., Bushnell W. R., Dik A. J., Carver T. L. W., editors. A Comprehensive Treatise. APS Press; 2002. [Google Scholar]
  • 7.Johnson R. Durable resistance: definition of, genetic control, and attainment in plant breeding. Phytopathology. 1981;71(6):567–568. doi: 10.1094/Phyto-71-567. [DOI] [Google Scholar]
  • 8.Niks R. E., Qi X., Marcel T. C. Quantitative resistance to biotrophic filamentous plant pathogens: concepts, misconceptions, and mechanisms. Annual Review of Phytopathology. 2015;53(1):445–470. doi: 10.1146/annurev-phyto-080614-115928. [DOI] [PubMed] [Google Scholar]
  • 9.Kuska M. T., Behmann J., Grosskinsky D. K., Roitsch T., Mahlein A. K. Screening of barley resistance against powdery mildew by simultaneous high-throughput enzyme activity signature profiling and multispectral imaging. Frontiers in Plant Science. 2018;9 doi: 10.3389/fpls.2018.01074. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Brugger A., Kuska M. T., Mahlein A.-K. Impact of compatible and incompatible barley—Blumeria graminis f.sp. hordei interactions on chlorophyll fluorescence parameters. Journal of Plant Diseases and Protection. 2017;125:177–186. doi: 10.1007/s41348-017-0129-1. [DOI] [Google Scholar]
  • 11.Wessling R., Panstruga R. Rapid quantification of plant-powdery mildew interactions by qPCR and conidiospore counts. Plant Methods. 2012;8(1):p. 35. doi: 10.1186/1746-4811-8-35. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Roitsch T., Cabrera-Bosquet L., Fournier A., et al. Review: new sensors and data-driven approaches-a path to next generation phenomics. Plant Science. 2019;282:2–10. doi: 10.1016/j.plantsci.2019.01.011. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Knauer U., Matros A., Petrovic T., Zanker T., Scott E. S., Seiffert U. Improved classification accuracy of powdery mildew infection levels of wine grapes by spatial-spectral analysis of hyperspectral images. Plant Methods. 2017;13(1) doi: 10.1186/s13007-017-0198-y. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Thomas S., Behmann J., Steier A., et al. Quantitative assessment of disease severity and rating of barley cultivars based on hyperspectral imaging in a non-invasive, automated phenotyping platform. Plant Methods. 2018;14(1) doi: 10.1186/s13007-018-0313-8. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Czedik-Eysenberg A., Seitner S., Güldener U., et al. The 'PhenoBox', a flexible, automated, open-source plant phenotyping solution. New Phytologist. 2018;219(2):808–823. doi: 10.1111/nph.15129. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Colmer J., O'Neill C. M., Wells R., et al. SeedGerm: a cost-effective phenotyping platform for automated seed imaging and machine-learning based phenotypic analysis of crop seed germination. New Phytologist. 2020;228(2):778–793. doi: 10.1111/nph.16736. [DOI] [PubMed] [Google Scholar]
  • 17.Lo Bianco M., Grillo O., Escobar Garcia P., Mascia F., Venora G., Bacchetta G. Morpho-colorimetric characterisation of Malva alliance taxa by seed image analysis. Plant Biology. 2017;19(1):90–98. doi: 10.1111/plb.12481. [DOI] [PubMed] [Google Scholar]
  • 18.Zhang S., Wang H., Huang W. Two-stage plant species recognition by local mean clustering and weighted sparse representation classification. Cluster Computing. 2017;20(2):1517–1525. doi: 10.1007/s10586-017-0859-7. [DOI] [Google Scholar]
  • 19.Jadhav S. B., Udupi V. R., Patil S. B. Identification of plant diseases using convolutional neural networks. International Journal of Information Technology. 2020 doi: 10.1007/s41870-020-00437-5. [DOI] [Google Scholar]
  • 20.Pound M. P., Atkinson J. A., Townsend A. J., et al. Deep machine learning provides state-of-the-art performance in image-based plant phenotyping. GigaScience. 2017;6(10) doi: 10.1093/gigascience/gix083. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Lin P., Li X. L., Chen Y. M., He Y. A deep convolutional neural network architecture for boosting image discrimination accuracy of rice species. Food and Bioprocess Technology. 2018;11(4):765–773. doi: 10.1007/s11947-017-2050-9. [DOI] [Google Scholar]
  • 22.Lee S. H., Chan C. S., Mayo S. J., Remagnino P. How deep learning extracts and learns leaf features for plant classification. Pattern Recognition. 2017;71:1–13. doi: 10.1016/j.patcog.2017.05.015. [DOI] [Google Scholar]
  • 23.Krizhevsky A., Sutskever I., Hinton G. E. ImageNet classification with deep convolutional neural networks. Communications of the ACM. 2017;60(6):84–90. doi: 10.1145/3065386. [DOI] [Google Scholar]
  • 24.Kaur S., Pandey S., Goel S. Plants disease identification and classification through leaf images: a survey. Archives of Computational Methods in Engineering. 2019;26(2):507–530. doi: 10.1007/s11831-018-9255-6. [DOI] [Google Scholar]
  • 25.He K., Gkioxari G., Dollár P., Girshick R. Mask R-CNN. 2017 IEEE International Conference on Computer Vision (ICCV); October 2017; Venice, Italy. pp. 2980–2988. [DOI] [Google Scholar]
  • 26.Ronneberger O., Fischer P., Brox T. In: Medical Image Computing and Computer-Assisted Intervention – MICCAI 2015. Navab N., Hornegger J., Wells W. M., Frangi A. F., editors. Cham: Springer International Publishing; 2015. [Google Scholar]
  • 27.Lin K., Gong L., Huang Y., Liu C., Pan J. Deep learning-based segmentation and quantification of cucumber powdery mildew using convolutional neural network. Frontiers in Plant Science. 2019;10 doi: 10.3389/fpls.2019.00155. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.Wang Q., Qi F., Sun M., Qu J., Xue J. Identification of tomato disease types and detection of infected areas based on deep convolutional neural networks and object detection techniques. Computational Intelligence and Neuroscience. 2019;2019:15. doi: 10.1155/2019/9142753.9142753 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29.Douchkov D. 2020. https://youtu.be/6GSfUzl1Txk.
  • 30.van der Walt S., Colbert S. C., Varoquaux G. The NumPy array: a structure for efficient numerical computation. Computing in Science & Engineering. 2011;13(2):22–30. doi: 10.1109/MCSE.2011.37. [DOI] [Google Scholar]
  • 31.Pedregosa F., Varoquaux G., Gramfort A., et al. Scikit-learn: machine learning in Python. Journal of Machine Learning Research. 2011;12:2825–2830. [Google Scholar]
  • 32.Lueck S., Beukert U., Douchkov D. BluVision Macro - a software for automated powdery mildew and rust disease quantification on detached leaves. Journal of Open Source Software. 2020;5(51, article 2259) doi: 10.21105/joss.02259. [DOI] [Google Scholar]
  • 33.Bates D., Mächler M., Bolker B., Walker S. Fitting linear mixed-effects models using lme4. Journal of Statistical Software. 2015;1(1) [Google Scholar]
  • 34.Mächler M., Bolker B., Walker S. Fitting linear mixed-effects models using ime4. Journal of Statistical Software. 2015;67(1):1–48. doi: 10.18637/jss.v067.i01. [DOI] [Google Scholar]
  • 35.Otsu N. A threshold selection method from gray-level histograms. IEEE Transactions on Systems, Man, and Cybernetics. 1979;9(1):62–66. doi: 10.1109/TSMC.1979.4310076. [DOI] [Google Scholar]
  • 36.Weisstein E. W. Weisstein E. W., editor. MathWorld - A Wolfram Web Resource. 2019. http://mathworld.wolfram.com/MooreNeighborhood.html.
  • 37.Novakovic J., Veljovic A. C-Support Vector Classification: selection of kernel and parameters in medical diagnosis. 2011 IEEE 9th International Symposium on Intelligent Systems and Informatics; 2011; Subotica, Serbia. pp. 465–470. [DOI] [Google Scholar]
  • 38.Cortes C., Vapnik V. Support-vector networks. Machine Learning. 1995;20(3):273–297. doi: 10.1007/BF00994018. [DOI] [Google Scholar]
  • 39.Ho T. K. Random decision forests. Proceedings of 3rd International Conference on Document Analysis and Recognition; 1995; Montreal, Quebec, Canada. pp. 278–282. [DOI] [Google Scholar]
  • 40.Ojala T., Pietikainen M., Harwood D. Performance evaluation of texture measures with classification based on Kullback discrimination of distributions. Proceedings of 12th International Conference on Pattern Recognition; 1994; Jerusalem, Israel. pp. 582–585. [DOI] [Google Scholar]
  • 41.Haralick R. M., Shanmugam K., Dinstein I.'. H. Textural features for image classification. IEEE Transactions on Systems, Man, and Cybernetics. 1973;SMC-3(6):610–621. doi: 10.1109/tsmc.1973.4309314. [DOI] [Google Scholar]
  • 42.Coelho L. P., Ahmed A., Arnold A., et al. Linking Literature, Information, and Knowledge for Biology. Vol. 6004. Berlin, Heidelberg: Springer; 2010. Structured Literature Image Finder: Extracting Information from Text and Images in Biomedical Literature; pp. 23–32. (Lecture Notes in Computer Science). [Google Scholar]
  • 43.Schweizer P., Pokorny J., Abderhalden O., Dudler R. A transient assay system for the functional assessment of defense-related genes in wheat. Molecular Plant-Microbe Interactions®. 1999;12(8):647–654. doi: 10.1094/MPMI.1999.12.8.647. [DOI] [PubMed] [Google Scholar]
  • 44.Nielsen K., Olsen O., Oliver R. A transient expression system to assay putative antifungal genes on powdery mildew infected barley leaves. Physiological and Molecular Plant Pathology. 1999;54(1-2):1–12. doi: 10.1006/pmpp.1998.0184. [DOI] [Google Scholar]
  • 45.Shirasu K., Nielsen K., Piffanelli P., Oliver R., Schulze-Lefert P. Cell-autonomous complementation of mlo resistance using a biolistic transient expression system. The Plant Journal. 1999;17(3):293–299. doi: 10.1046/j.1365-313X.1999.00376.x. [DOI] [Google Scholar]
  • 46.Wilkinson M. D., Dumontier M., Aalbersberg I. J. J., et al. The FAIR Guiding Principles for scientific data management and stewardship. Scientific Data. 2016;3(1) doi: 10.1038/sdata.2016.18. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 47.König P., Beier S., Basterrechea M., et al. BRIDGE – a visual analytics web tool for barley genebank genomics. Frontiers in Plant Science. 2020;11 doi: 10.3389/fpls.2020.00701. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 48.Douchkov D., Lueck S. e!DAL - Plant Genomics & Phenomics Research Data Repository. 2019. [DOI]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Supplementary Materials

Supplemental Figure S1: schematic drawing of the image acquisition hardware of the Macrophenomics module Macrobot 2.0 (top and side views). Supplemental Table S1: validation experiment, all data (Excel sheet). Supplemental Table S2: values in Figure 4(a)—evaluation of different classifiers on HSV_H_channel (5000 pixels per class, n = 10). Supplemental Table S3: values in Figure 4(b)—evaluation of the random forest classifier with a different number of trees on HSV_H_channel (5000 pixels per class, n = 10). Supplemental Table S4: values in Figure 5(a)—evaluation of different color pixel classification methods (n = 10). Supplemental Table S5: values in Figure 5(b)—evaluation of texture features (n = 10). Supplemental Table S6: list of wheat genotypes (Excel sheet).

Data Availability Statement

Image data used for validation of the Macrobot algorithm is available at [48]. The image analysis software is available at [32].


Articles from Plant Phenomics are provided here courtesy of American Association for the Advancement of Science (AAAS) and Nanjing Agricultural University

RESOURCES