Skip to main content
Plant Methods logoLink to Plant Methods
. 2017 Jan 31;13:7. doi: 10.1186/s13007-017-0157-7

A high-throughput stereo-imaging system for quantifying rape leaf traits during the seedling stage

Xiong Xiong 1,#, Lejun Yu 1,#, Wanneng Yang 2,3, Meng Liu 2, Ni Jiang 1, Di Wu 3, Guoxing Chen 4, Lizhong Xiong 2, Kede Liu 2, Qian Liu 1,
PMCID: PMC5282657  PMID: 28163771

Abstract

Background

The fitness of the rape leaf is closely related to its biomass and photosynthesis. The study of leaf traits is significant for improving rape leaf production and optimizing crop management. Canopy structure and individual leaf traits are the major indicators of quality during the rape seedling stage. Differences in canopy structure reflect the influence of environmental factors such as water, sunlight and nutrient supply. The traits of individual rape leaves traits indicate the growth period of the rape as well as its canopy shape.

Results

We established a high-throughput stereo-imaging system for the reconstruction of the three-dimensional canopy structure of rape seedlings from which leaf area and plant height can be extracted. To evaluate the measurement accuracy of leaf area and plant height, 66 rape seedlings were randomly selected for automatic and destructive measurements. Compared with the manual measurements, the mean absolute percentage error of automatic leaf area and plant height measurements was 3.68 and 6.18%, respectively, and the squares of the correlation coefficients (R2) were 0.984 and 0.845, respectively. Compared with the two-dimensional projective imaging method, the leaf area extracted using stereo-imaging was more accurate. In addition, a semi-automatic image analysis pipeline was developed to extract 19 individual leaf shape traits, including 11 scale-invariant traits, 3 inner cavity related traits, and 5 margin-related traits, from the images acquired by the stereo-imaging system. We used these quantified traits to classify rapes according to three different leaf shapes: mosaic-leaf, semi-mosaic-leaf, and round-leaf. Based on testing of 801 seedling rape samples, we found that the leave-one-out cross validation classification accuracy was 94.4, 95.6, and 94.8% for stepwise discriminant analysis, the support vector machine method and the random forest method, respectively.

Conclusions

In this study, a nondestructive and high-throughput stereo-imaging system was developed to quantify canopy three-dimensional structure and individual leaf shape traits with improved accuracy, with implications for rape phenotyping, functional genomics, and breeding.

Electronic supplementary material

The online version of this article (doi:10.1186/s13007-017-0157-7) contains supplementary material, which is available to authorized users.

Keywords: Stereo-imaging system, Canopy three-dimensional structure, Individual leaf traits, Morphological classification

Background

Oilseed rape (Brassica napus) is an important species that is cultivated in many countries for its valuable oil and protein [15]. The area planted with oilseed rape has rapidly increased in recent decades [6]. The leaf is of fundamental importance to the rape, acting as the power generator and aerial environmental sensor of the plant [7, 8]. Leaves are primarily involved in photosynthesis and transpiration, thereby influencing crop yield [9, 10]. The size, shape, area and number of leaves are of great significance to plant science, allowing scientists to distinguish between different species and even to model climate change [11]. Moreover, plant canopy architecture is of major interest for plant phenotyping. Variations in canopy structure have been linked to canopy function and have been attributed to genetic variability as well as a reaction to environmental factors such as light, water, and nutrient supplies as well as stress [12]. Thus, canopy structure is an essential variable for plant’s adaptation to its environment [13, 14]. It is therefore important to study the oilseed rape phenotypic traits of both individual leaf shape and plant canopy structure.

Many researchers have carried out studies of individual leaf traits [1519]. In many cases, rapeseed species can be distinguished by aspects of leaf shape, flower shape, or branching structure. Shape is, of course, important in many other disciplines [11]. To characterize these properties, O’Neal et al. [20] applied a desk-top scanner and public domain software to extract individual leaf shape traits, including leaf height, leaf width. However, the efficiency of this process is problematic: each leaf must be removed from the plant and scanned into a digital format. In addition, some complex traits such as leaf serration and leaf margin can’t be assessed by this method. In an attempt to measure leaf area easier and more accurate, the new software namely “Compu Eye, Leaf & Symptom Area” was developed by Bakr et al. [21], etc. The purpose of this software is to obtain the symptom area for each leaf. But, this method offers no method to quantify leaf serration and morphology traits. Thus, this software has some limitations in practical. Igathinathane et al. [22] designed software that uses the computer monitor as the working surface to trace leaf outline and determines leaf area, perimeter, length, and width. This software offers no method to quantify leaf serration and inner cavity-related traits. Also, this is a semi-automatic program and the interactive processes are complex and tedious. Weight et al. [23] reported the development of LeafAnalyser, which is an excellent tool to facilitate PCA analysis of leaf shape parameters. However, the leaf petiole region did not remove when analyzing the leaf traits and the software was not released as open source, negating the possibility of further development by the community. Bylesjö et al. [7] designed a tool to extract classical indicators of blade dimensions and leaf area, as well as measurements that indicate asymmetry in leaf shape and leaf serration traits. This software not only obtains object boundaries but also analyzes serration traits. However, it requires the user to analyze leaves in vitro and to correctly characterize the blade azimuth for subsequent image analysis, which limits the throughput of the analysis. Dengkui et al. [24] designed a tool to acquire plant growth information by abstracting the plant morphological characters, size and color of leaves, etc. The morphological operation has been used to remove petiole, which will influence the accuracy of leaf margin information extraction. But, the method offers no method to quantify leaf serration and inner cavity-related traits. Yang et al. [25] designed a device “HLS” for assessing leaf number, area, and shape. The device is efficient and can process multiple blades in parallel. However, all blades must be cut from the plant before insertion into the HLS device. Furthermore, the present equipment for extracting serrated blade edge traits is insufficient.

The work described above focuses on the morphological traits of individual leaves. However, three-dimensional canopy structure also plays an important role in sustaining plant function. The canopy structure contains useful information regarding developmental stage during the vegetation period as well as yield-forming parameters [26, 27]. Three-dimensional imaging methods can be broadly classified into two types: active and passive [14]. Commonly used active light projection technologies include laser scanning and structured light. Light detection and ranging (LIDAR) laser scanners have emerged as a powerful active sensing tool for direct three-dimensional measurement of plant height, canopy structure, plant growth, and shape responses [28]. The precision of laser scanner systems is very high, but the scanning time is very long, reducing the system’s throughput. In structure light systems, the Kinect Microsoft RGB-depth camera [29] is used as a depth camera to shine light onto the object scene. The light reflected from the scene is used to build the depth image by measuring the deformation of the spatially structured lighting pattern [30]. The system produces 640 × 480 pixels RGB-depth images coded with a 16-bits dynamic that are acquired at a rate of 30 frames per second [13]. The imaging speed of these systems is very high, nearly satisfying the demands of real-time measurement. However, the measurement accuracy exhibits low spatial resolution in comparison with a standard RGB camera. One problem with laser-based and structured light systems is that they do not work well with reflective objects, and it is often necessary to coat the surface with a non-reflective layer that can lead to the collection of unsatisfactory texture data [31]. In addition, the method used for volumetric reconstruction from multiple images has been proposed to be a passive three-dimensional imaging technology [3234]. It works by obtaining multiple images from different fixed angles. Here, a rotated plate is used to achieve multi-angle imaging, which will result in time-consuming rotations. Moreover, this method requires a significant amount of post-processing. Also, binocular/multi-view stereo imaging approach is another major passive three-dimensional imaging technology [35, 36]. There are some applications of using binocular/multi-view stereo vision for plant sensing. For automatic robot or vehicle-mounted system, the binocular stereo system is a common component for obtaining distance depth information or field plant 3D structure [3740]. Moreover, the binocular stereo is also used in small- to medium-sized plant canopies reconstruction. Ivanov et al. [41] applied film-based stereo photogrammetry to reconstruct the maize canopy, where the plant canopy geometrical structure was analyzed and different simulation procedures were carried out to analyze leaf position and orientation and leaf area distribution. Andersen et al. [42] designed simulated annealing (SA) binocular stereo match algorithm for young wheat plants and analyzed height and total leaf area for single wheat plant. Biskup et al. [43] designed a stereo vision system with two cameras to build 3D models of soybean canopy and analyzed the angle of inclination of the leaves. Also, for isolated leaf, Biskup established a stereoscopic imaging system, which quantifies surface growth of isolated leaf discs floating on nutrient solution in wells of microtiter plates [44]. Müller-Linow et al. [12] developed a software package, which provides tools for the quantification of leaf surface properties within natural canopies via 3-D reconstruction from binocular stereo images. Furthermore, the multi-view stereo 3D reconstruction for plant phenotyping is also widely used combining with SfM- and MVS-based photogrammetric method. Lou et al. [45] described an accurate multi-view stereo (MVS) 3D reconstruction method of plants using multi-view images, which takes both accuracy and efficiency into account. Several plants, including arabidopsis, wheat and maize, are used to evaluate the performance of reconstruction algorithm. Rose et al. [46] developed a multi-view stereo system to evaluate the potential measuring accuracy of a SfM- and MVS-based photogrammetric method for the task of organ-level tomato plant phenotyping. The leaf area, main stem height and convex hull of the complete tomato plant are analyzed. Miller et al. [47] applied a low-cost hand-held camera to accurately extract height, crown spread, crown depth, stem diameter and volume of small potted trees. The multi-view stereo-photogrammetry was used to generate 3D point clouds. From the literatures above, the binocular stereo is usually used in small-sized plant canopies reconstruction by using two top-view cameras and the multi-view stereo 3D reconstruction method is applied for organ-level plant 3D phenotyping.

In this study, we attempt to create a three-dimensional surface model of the rape canopy from images taken by double top-view cameras, and we estimate geometric attributes such as plant height and canopy leaf area. For RGB images collected using a stereo-imaging system, a novel image analysis pipeline for the accurate quantification seedling rape leaf traits was developed. We are thus able to perform leaf shape analysis, including contour signatures and shape features.

Results and discussion

Development of a stereo-imaging system

In order to extract canopy leaf area, plant height and canopy three-dimensional structure, we developed a stereo-imaging system consisting of three major units: an imaging unit, a transportation unit and a control unit (Fig. 1a). For the imaging unit, we utilized two identical RGB cameras [AVT Stingray F-504B/F-504C, Allied Vision Technologies Corporation, 2452 (H) × 2056 (V) resolution] with 8 mm fixed focal lenses (M1214-MP, Computar Corporation), two LED lamps and a lifting platform. The RGB cameras are fixed to ensure that the two main optical axes are parallel and that the two imaging planes are located at the same horizontal level. An automatic trigger acquires image pairs, and software was developed to obtain the color image pairs simultaneously. The lifting platform can be used to adjust the imaging region [537.5 mm (H) × 449.9 mm (V)] and spatial resolution (0.2188 mm/pixel). In addition, the computer workstation (HP xw6400, Hewlett-Packard Development Company, USA) plays the role of central control unit, and utilizes software developed by LabVIEW 8.6 (Nation Instruments, USA) to communicate with the two RGB cameras. In order to achieve high-throughput measurements, the stereo-imaging system was integrated into an automated high-throughput phenotyping facility developed in our previous work [48]. Two optional processing modes were developed to reconstruct the three-dimensional structure of the seedling rape canopy (Fig. 1b) and to extract individual rape leaf traits (Fig. 1c).

Fig. 1.

Fig. 1

Development of stereo imaging system and two optional processing modes. a The inspection unit. b The three-dimensional structure of seedling rape canopy. We can extract plant height and canopy leaf area from the three-dimensional structure. c The individual rape leaf morphological traits, including scale-invariant shape traits, cavity-related traits and margin-related traits

Three-dimensional structure of the seedling rape canopy

Plant canopy structure can be described by a range of complex and variable phenotypic traits that dictate the function of plant [49]. Here, canopy three–dimensional point cloud data were extracted from pairs of digital color images obtained under a constant light environment. The point cloud size for each canopy reconstruction is nearly 5.5 Mb. The user-friendly software interface for three-dimensional reconstruction is shown in Additional file 1: Figure S3, and the final reconstructed seedling rape canopy shown in Fig. 2 from three different perspectives.

Fig. 2.

Fig. 2

Three-dimensional reconstructions for seedling rape canopy at three different perspectives. a The original two-dimensional rape leaf image. bd There are three types of perspectives for rape three-dimensional canopy structure

Leaf area and plant height

From the generated three-dimensional structure, we were able to extract two important parameters: leaf area and plant height. To evaluate leaf area based on canopy level, the Delaunay algorithm was used in the process of three-dimensional mesh generation. Figure 3 shows the detailed process for leaf triangular patches generation. After the stereo-imaging for seedling rape canopy, a set of 3D point cloud can be obtained (Fig. 3a). Color differences in point cloud represents different rape leaf. The matlab functions “trimesh” and “delaunay” based on Lifting Method [50] are applied to achieve Delaunay algorithm. The “delaunay” function produces an isolated triangulation, which is useful for applications like plotting surfaces via the “trimesh” function. The stack of triangular patches forms the 3D leaf region and a smoothing mechanism is used to extract smooth triangular mesh (Fig. 3b, c). Finally, we can obtain the Delaunay triangulations in rape canopy level and the sum area of all smooth triangular meshes is the canopy leaf area (Fig. 3d).

Fig. 3.

Fig. 3

The schematic diagram for three-dimensional reconstruction based on Delaunay triangulations. a The three-dimensional scatter plot for oilseed rape canopy leaf (take an oilseed rape with four leaves, for example). Color differences in point cloud represents different rape leaf. b The reconstructed oilseed leaf [the rectangular area in image (a)] based on Delaunay triangulations. Color differences in smooth triangular mesh reflect the depth information in different leaf region. c The left part is the smooth 3D triangular mesh, which is projected to xy plane. The triangular patches image in the lower-right corner shows the local details in the red rectangle. The right part is the illustration for left oilseed rape leaf in three-dimensional space. d The final result for Delaunay triangulations reconstruction in canopy level. Color differences in smooth triangular mesh reflect the depth information in different leaf region

To evaluate the measurement accuracy of leaf area and plant height (vertical distance from the edge of a plastic pot to the tip of longest leaf), 66 rape seedling plants were randomly selected for manual measurement. Figure 4 shows the results of manual observation versus automatic observation. The MAPE values were 3.68% for leaf area and 6.18% for plant height, and the squares of the correlation coefficients (R 2) were 0.984 and 0.845, respectively. Detailed experimental data are presented in Additional file 2. The result shows that the stereo-vision method has a good potential for accurate measurement.

Fig. 4.

Fig. 4

The performance evaluation of the leaf area and plant height. a The scatter plots of the stereo imaging measurement versus the HLS measurement for leaf area. b The scatter plots of the stereo imaging measurement versus the manual measurement by a ruler for calculating the plant height

Individual leaf traits

Shape-based individual leaf traits, such as leaf size, vein network and leaf margin, are currently used for plant species identification and quantitative trait loci mapping [11, 16, 51, 52]. These shape-based morphological traits can be extracted using our image analysis pipeline. However, these traits alone do not reflect differences of leave shape due to variation in leaf size. We therefore must consider several new characteristic parameters that are not influenced by leaf size. Here, 19 shape related traits, including 11 scale-invariant traits, 3 inner cavity-related traits, and 5 margin-related traits, are proposed (Additional file 3). The definitions of all shape-related traits are shown in Fig. 5 and Table 1, and the computational formulas are provided in Eqs. 19.

AspectRatio=LengthboundingboxWidthboundingbox 1
Rectangularity=AreaobjectAreaboundingbox 2
AreaConvexity=AreaobjectAreaconvexhull 3
PerimeterConvexity=PerimeterobjectPerimeterconvexhull 4
Sphericity=4π×AreaobjectPerimeterconvexhull2 5
Eccentricity=AxisLengthlongAxisLengthshort 6
FormFactor=4π×AreaobjectPerimeterobject2 7
P/A=PerimeterobjectAreaobject 8
Circularity=RadiusinscribedcircleRadiusexcircle 9

Fig. 5.

Fig. 5

The extraction of individual leaf traits. a The extraction of the inscribed circle, circumscribed circle and minimum circumscribed box. b Analysis of the inner edges and cavities. c Calculate the leaf margin and indents by using the convex hull algorithm

Table 1.

Definitions of nineteen leaf shape-related traits

Classification Variable Definition
Scale-invariant traits AA The aspect ratio of leaf minimum circumscribed box
R The ratio of leaf area to minimum circumscribed box area
AC The ratio of leaf area to leaf convex hull area
PC The ratio of leaf circumference to leaf convex hull perimeter
S The ratio of leaf area to the square of leaf convex hull perimeter
E The ratio of long axis of ellipse to short axis of ellipse
FF The ratio of leaf area to the square of leaf perimeter
PAR The ratio of leaf circumference to leaf area
SFD Reflect the effectiveness of the occupies space without image cropping [53]
IFD Reflect the effectiveness of the occupies space with image cropping [53]
C The ratio of inscribed circle radius to circumscribed circle radius
Cavity traits NIC The number of inner cavities
APIC The average perimeter of inner cavities
AAIC The average area of inner cavities
Margin related traits TNI Total number of indents
ENI Effective number of indents
ADI The average depth of indents
ADEI The average depth of effective indents
AVE The average calculated value of effectiveness

As seen in Fig. 5a, the red and purple circles represent the inscribed circle and circumscribed circle, respectively. The yellow ellipse indicates the result of elliptical fitting for a rape leaf. The green rectangle represents the minimum circumscribed box of the blade region. The inner cavities are by definition surrounded by a boundary region that is not connected to the outer boundary of the object. As seen in Fig. 5b, the green regions delineate inner cavities. In addition, the red lines (Fig. 5c), indicate the outer boundary of the individual leaf, while the blue lines demarcate the convex hull. The green points on the convex hull lines represent the serration points, which reflect to the vertices in all directions. The intermediate region between two serration points defines an indent. For each indent region, two serration points can be connected by a straight line, and the depth of the indent is measured as the longest distance from an indent point to the corresponding straight line. The effectiveness of the indents is calculated using the following Eq. 10. When the ratio is greater than 0.3, we consider the indents to be effective indents.

iseffective=depthmin(height,width)/2 10

where, height and width represent the number of minimum circumscribed box rows and cols, respectively. In addition, depth represents the distance from the indent point to the corresponding convex hull straight line.

The software interface for extracting individual leaf traits is shown in Additional file 4: Figure S4.

Stepwise discriminant analysis

801 samples were randomly selected and divided into two groups: 402 samples with three different leaf shapes (mosaic-leaf, semi-mosaic-leaf, and round-leaf) comprising the training group and 399 samples with three different leaf shapes comprising the testing group. According to our classification, eleven of nineteen significant traits were selected by stepwise discriminant analysis as independent variables to construct two decision functions. The final classification using the two decision functions is shown in Fig. 6.

Fig. 6.

Fig. 6

Stepwise discriminant analysis classification results. The abscissa and ordinate represent two classification functions, which were built by stepwise discriminant analysis algorithm. The black squares reflect the center of different groups, which can be calculated with two decision functions. The red, green and blue points represent round-leaf, semi-mosaic-leaf and mosaic-leaf, respectively

The black square blocks represent the center of the three different leaf shapes, which can be calculated using two decision functions. In Fig. 6 (red points mean round-leaf, green points mean semi-mosaic-leaf and blue points mean mosaic-leaf) and Table 2, the classification accuracy is given. We found that 92.7% of the tested grouped cases were correctly classified. In addition, we applied the leave-one-out cross-validation (LOO-CV) [54] to assess the accuracy of the classification model and found that 94.4% of the cross-validated grouped cases were correctly classified.

Table 2.

Stepwise discriminant analysis classification results

Leaf type Predicted group membership Total
Mosaic Semi-mosaic Round
Trained group % Mosaic-leaf 98.3 1.7 0 100.0
Semi-mosaic-leaf 1.4 90.3 8.3 100.0
Round-leaf 0 0 100.0 100.0
Tested group % Mosaic-leaf 97.5 2.5 0 100.0
Semi-mosaic-leaf 7.6 88.9 3.5 100.0
Round-leaf 0 7.4 92.6 100.0
Cross-validateda % Mosaic-leaf 95.9 4.1 0 100.0
Semi-mosaic-leaf 2.1 89.3 8.7 100.0
Round-leaf 0 1.5 98.5 100.0

96.0% of trained grouped cases correctly classified

92.7% of tested grouped cases correctly classified

94.4% of cross-validated grouped cases correctly classified

aCross validation is conducted only for those cases in the analysis. In cross validation, each case is classified by the functions derived from all cases other than that case

Stepwise discriminant analysis (SDA) [55] has been proven to effectively classify different rape leaf shapes. Moreover, in stepwise discriminant analysis, the first selected variable carries more weight in classification. In this study, the first four selected variables—form factor (FF), area convexity (AC), the average depth of effective indents (ADEI), and perimeter convexity (PC)—represent nearly 97% of the classification ability, as shown in Fig. 7. As we can see from Fig. 7, introducing a new variable will have little impact on the classification results after the selection of the first four variables. For FF, the round-leaf always has small perimeter for a given area, so the FF for the round-leaf shape is smaller than that of the mosaic-leaf and semi-mosaic-leaf. AC represents the ratio of the leaf area to the leaf convex hull area, which is an important parameter that reflects leaf morphology. For mosaic-leaf, the serrate border feature increases the convex area of the leaf. Thus, the duty ratio relative to its convex hull decreases markedly in comparison with the two other leaf shapes. Due to its serrated edge, the mosaic-leaf exhibits deeper indents compared with other two shapes.

Fig. 7.

Fig. 7

The performance analysis of traits in stepwise discriminant analysis. The screening results of the stepwise discriminant analysis. The traits from top to bottom reflect the order of characteristics screening. The value of Wilks’ Lambda statistic represents the discriminant ability after entering current traits

Support vector machine (SVM)

Support vector machine (SVM) is a standard classification technique that has been shown to produce state-of-the-art results in many classification problems [54, 56]. To apply SVM to our rape seedling leaf classification, nearly half of the 801 rape samples (402 samples) were used as training parameters and the other samples (399 samples) without labels were divided into a testing group for comparison of the results. We found that 94.7% of the tested grouped cases were correctly classified. In addition, the leave-one-out (LOO-CV) accuracy is 95.6% for the cross-validated group. The specific classification accuracy is shown in Table 3.

Table 3.

SVM classification results

Leaf type Predicted group membership Total
Mosaic Semi-mosaic Round
Trained group % Mosaic-leaf 100.0 0 0 100.0
Semi-mosaic-leaf 0 99.3 0.7 100.0
Round-leaf 0 0 100.0 100.0
Tested group % Mosaic-leaf 100.0 0 0 100.0
Semi-mosaic-leaf 6.2 92.4 1.4 100.0
Round-leaf 0 7.4 92.6 100.0
Cross-validateda % Mosaic-leaf 96.7 3.3 0 100.0
Semi-mosaic-leaf 3.1 93.1 3.8 100.0
Round-leaf 0 2.6 97.4 100.0

99.8% of trained grouped cases correctly classified

94.7% of tested grouped cases correctly classified

95.6% of cross-validated grouped cases correctly classified

aCross validation is conducted only for those cases in the analysis. In cross validation, each case is classified by the functions derived from all cases other than that case

Random forest

In essence, the random forest (RF) model is a multiple decision trees classifier, and it is widely used in regression analysis and multi-classification [57, 58]. In this study, 402 samples with three different shapes were used to construct random forest model. The other 399 samples with category labels as testing group were applied to evaluate the performance of classification. The final classification accuracy for a test group is 91.7%, and the leave-one-out cross-validation accuracy is 94.8%. The specific classification accuracy is shown in Table 4.

Table 4.

Random forest classification results

Leaf type Predicted group membership Total
Mosaic Semi-mosaic Round
Trained group % Mosaic-leaf 100.0 0 0 100.0
Semi-mosaic-leaf 0 100.0 0 100.0
Round-leaf 0 0 100.0 100.0
Tested group % Mosaic-leaf 99.2 0.8 0 100.0
Semi-mosaic-leaf 8.3 87.5 4.2 100.0
Round-leaf 0 10.4 89.6 100.0
Cross-validateda % Mosaic-leaf 96.7 3.3 0 100.0
Semi-mosaic-leaf 3.5 91.0 5.5 100.0
Round-leaf 0 3.0 97.0 100.0

100.0% of trained grouped cases correctly classified

91.7% of tested grouped cases correctly classified

94.8% of cross-validated grouped cases correctly classified

aCross validation is conducted only for those cases in the analysis. In cross validation, each case is classified by the functions derived from all cases other than that case

A comparison of the performance of the three methods of classification

As shown in Table 5, all three methods were able to satisfactorily classify leaves. The final classification accuracy for test group (399 samples) is 92.7, 94.7, 91.7% for stepwise discriminant analysis (SDA), support vector machine (SVM) and random forests (RF), respectively. In addition, the leave-one-out cross validation classification accuracy is 94.4% for SDA, 95.6% for SVM and 94.8% for RF algorithm. Among them, the stepwise discriminant analysis has a better prediction effect on round-leaf, while the support vector machine classifier is the most sensitive to mosaic-leaf. From the perspective of predicated group accuracy and the leave-one-out cross-validated results, the most reliable forecasting model was established by SVM algorithm.

Table 5.

A comparison of three classification methods

Methods of classification Predicted group accuracy Leave-one-out cross-validated (LOO-CV)
Mosaic Semi-mosaic Round Total (n = 399)
Stepwise discriminant analysis 98.3% 90.3% 100.0% 92.7% 94.4%
Support vector machine (SVM) 100% 92.4% 92.6% 94.7% 95.6%
Random forests classifier 99.2% 87.5% 89.6% 91.7% 94.8%

Fixed half selected for training (402), and the other half (399) for testing

The performance of efficiency and accuracy

In this work, the stereo-imaging system is integrated to the high-throughput phenotyping facility. Each pot-grown rape would be transported by the conveyor, and the image pairs were acquired by the two top-view cameras at the same time. The inspection procedure is fully automated and highly efficient (45 s per plant) [48]. All the image processing works are carried out after the completion of image acquiring. Here, the time for image processing consists of two parts: canopy 3D reconstruction and individual leaf traits extraction. Usually, the time for canopy three-dimensional reconstruction is closely linked to the size of oilseed rape. After evaluated with 10 different size seedling rape samples, the average processing time for each canopy 3D reconstruction and data extraction of leaf area and plant height is about 43.46 s; for manual interaction, the time for each individual leaf extraction is about 8 s. The detailed description for manual part is shown in video (Additional file 5). Moreover, the two independent parts could run in parallel to save time. In this way, the total processing time depends on the longer part. So, the manual interaction does not lag the efficiency of the high-throughput platform. In addition, the manual interaction method can extract individual leaf with more accuracy compared with automatic segmentation, and the efficiency is also satisfactory. Assuming that the system can work 8 h a day, then, about 660 pots can handle just one day, which is an acceptable number for high-throughput.

From the view of measurement accuracy for rape seedling leaf area, a comparison result of two different methods is shown in Fig. 8a. The red scatter points represent the leaf area result of two-dimensional projective method by only use one top-view image. After the rape was segmented from the background in the top-view image, the area of each rape is calculated by multiplying the pixel area and average spatial resolution, while blue scatter points indicate the result of three-dimensional stereo measurement by analyzing canopy structure. The MAPE values were 3.68% for three-dimensional stereo measurement and 11.44% for two-dimensional projective measurement, and the square of correlation coefficients (R 2) for three- and two-dimensional measurements was 0.984 and 0.938, respectively. Obviously, compared with two-dimensional imaging, the stereo measurement considering more spatial information of rape leaf can indicate more accuracy of leaf area in real world. Moreover, the area errors are almost below eight percent by using the stereo measurement method (Fig. 8b).

Fig. 8.

Fig. 8

The comparison of two different methods for leaf area. a The red scatter points represent the leaf area result of the two-dimensional projective method by only using one top image. The blue scatter points indicate the result of three-dimensional measurement by using stereo imaging algorithm for leaf area. b The distribution of percentage error with two different methods

The overlapping situation in stereo-imaging

The overlap of oilseed rape leaves is surely a difficult issue for binocular stereo-imaging. In this study, the oilseed rapes are at the seedling stage, which have less overlapping situation. Actually, for some situation (round-leaf), the overlap can be solved and recovered. The following part only considers the round-leaf (Fig. 9a). The detailed implementation steps are as follows: In the first step, we need to segment the overlapped leaf binary image. Secondly, the contour of overlapped leaf is extracted by using the front binary image. Then, the polygonal approximation [59] is used to represent the overlapped contour. This is an important step to trim away the small-scale rough fluctuations. Next, we need to detect the concave points [60] and segment the polygonal contour (Fig. 9b). Finally, the ellipse fitting [61] is chosen to recover the overlapped leaf region for round-leaf (Fig. 9c). The detailed algorithm description can refer to Additional files 6 and Additional File 7: Figure S5. The key for above algorithm is based on a priori knowledge: the oilseed rape leaf is approximate circle. Thus, for mosaic-leaf and semi-mosaic-leaf, the above method is useless.

Fig. 9.

Fig. 9

Recovering the overlapped leaf region for round-leaf. a The original overlapped round-leaf. b The concave points (red) and polygonal boundary (white) of overlapped round-leaf. c The ellipse fitting is chosen to recover the overlapped leaf region for round-leaf

Conclusion

In this study, we establish a nondestructive and high-throughput stereo-imaging system for screening leaf canopy three-dimensional structure and individual leaf phenotypic traits. Compared with manual measurements, the squares of the correlation coefficients (R 2) for leaf area and plant height are 0.984 and 0.845, respectively. Moreover, 19 morphological traits were applied in morphology classification of three different rape leaf shapes. Three classifiers (SDA, SVM, and RF) were used and compared, and the better classification accuracy with SVM is 94.7% for 399 test samples. In conclusion, we developed a high-throughput stereo-imaging system to quantify leaf area, plant height, and leaf shape with more accuracy, which will benefit rape phenotyping, functional genomics, and breeding.

Methods

Plant materials and measurements

In total, 801 Brassica napus with three different shapes, including mosaic-leaf, semi-mosaic-leaf and round-leaf (Additional file 8: Figure S2), were analyzed in this study. Seeds were sown and germinated, and plants were grown up to the seedling stage. All plants were cultivated in plastic pots of 23.5 cm diameter with approximately 6 L of experimental soil. All pots were randomly distributed over a glasshouse compartment to control the growth conditions. Approximately 30 days after sowing, three experienced agronomists recorded the leaf shape using the visual method. The final statistical classification result would abide by the majority rule. All the experimental samples were measured with our stereo-imaging system to obtain image pairs. Among them, 66 rape plants were randomly selected to reconstruct the canopy three-dimensional structure, extract leaf area and calculate plant height. To estimate the accuracy of measurement, the plant leaf area was measured with the HLS [25] and plant height was measured manually by well-trained worker. In order to evaluate the extraction of individual leaf traits, a biological classification for three different leaf shapes was proposed. All samples were divided into two groups: one group consists of 402 samples with three labels (mosaic-leaf, semi-mosaic-leaf and round-leaf) as the training group. The other group consists of 399 samples without labels as the testing group. All the training group samples are selected randomly to balance the number of different shapes.

Image analysis for canopy three-dimensional reconstruction

The main content of this part is to describe the specific image processing steps of canopy 3D reconstruction for seedling rape. The first step is camera stereo calibration. To achieve this step, a black and white calibration pattern [31] pasted on a plastic plate was used to obtain 20–25 image pairs. To ensure the accuracy of the calibration, the imaging angles should have obvious differences. For original image pairs (Fig. 10a), the corresponding feature in the left and right original image is not on the same horizontal baseline. Here, Bouguet algorithm [62] was used to rectify two original images. The final rectified image pairs were shown in Fig. 10b. Considering the influence of environmental light, an automatically segmenting method [63], adopts normalized RGB component to get binary image of blade region (Fig. 10c). Considering the slender characteristics of the stem, the morphological opening operation is used to remove the stem, and connected component mark technology is used to distinguish different leaves (Fig. 10d). The next work is to match the corresponding feature points in the left and right rectified images. Here, the library for efficient large-scale stereo matching [64] is used to compute the left and right disparity map. In the actual situation, two thorny situations might happen [65, 66]. The first one is mismatch, which means that there are no matching pixels or wrong matching pixels. The second thorny situation is occlusion, which means that some pixels appears only in an image, and can’t see in another image (Additional file 9: Figure S6). If we don’t take some special measures to focus on region where the mismatched and occluded situations are serious, there will have some wrong in the process of 3D point clouds extraction. So, it is important to rectify the disparity map. The specific process is described in Additional file 10 and the rectified left disparity image is shown in Fig. 10e. According to the principle of triangular range (Additional file 11: Figure S1), we can extract the three-dimensional point cloud data of the canopy leaves. After removing the isolate points, triangle patches are used as the surface of canopy leaves by using Delaunay triangulation algorithm [50]. The final result of canopy reconstruction is shown in Fig. 2b–d. Detailed processing for canopy 3D reconstruction and triangle patches generation has been described in Additional file 10 and Fig. 3. The code for Delaunay algorithm is shown in Additional file 12.

Fig. 10.

Fig. 10

The primary image analysis procedures for seedling rape canopy three-dimensional reconstruction. a The same feature region in the left and right original image was not on the same horizontal baseline. b The same feature region in the left and right rectified image was at the same horizontal baseline. c The segmentation of rape leaf region. d The stems are removed, and the leaf region is marked. e A local rectification for discontinuous regions and false matching regions. f The main software flow chart

Image analysis for individual leaf

The main content of this part is to describe the specific image processing steps for individual leaf extraction. Firstly, the user needs to click the left mouse button and drag it to choose a rectangular box. In this rectangular box, the individual leaf must be typical and representative (Fig. 11a). All selected rectangular images are saved in PNG format for subsequent analysis. Usually, the selected individual leaf has a long petiole part, the existence of which will seriously impact the analysis of blade traits. So, the next step is to remove petiole. The difference between blade and petiole in color and texture is tiny. But from the view of shape, petiole region is more slender than blade. With that mechanism we can remove petiole. The detailed operating processing includes the following steps: (1) Marking two points on the petiole (Fig. 11b) and rotating the rectangular image so that the direction of the petiole is downward (Fig. 11c). (2) Segmenting rotated rectangular image to obtain binary leaf image (Fig. 11d). Here, the excess green vegetation (ExG) [67] and excess red vegetation (ExR) indices [68] were used to extract binary leaf image. (3) From the bottom to top search binary image to remove the pixel width less than a specified threshold area (Fig. 11e). Here, the area threshold is set to 25, which is an appropriate value determined by lots of preliminary experiments. After removing the petiole, the next step is to remove connected components that were erroneously selected. The situation that other partial leaf region might be chosen in the rectangular region was always happened. Usually, the target leaf region had the largest area. So, only thing we need to do is to keep the largest connected component as the target individual leaf. In addition, because of the binarization segmentation error, small holes might appear on the target blade region (the red square in Fig. 11e). Usually, these holes have a small number of pixels compared with other real holes. An area threshold is used to fill the small area holes. The final individual leaf is shown in Fig. 11f. The detailed processing flow and computational formulas are in Additional file 13.

Fig. 11.

Fig. 11

The primary procedure for individual leaf traits extract. a The original rape leaf image. b The selected single rape leaf region. c Select two points on the image. The first point is the intersective region of stem and leaf and the second point is used to decide the direction of rotation. d The rotated oilseed rape leaf image. e Using normalized EG and ER to segment the leaf region. f Remove the petiole region and connected region screening

Three different classification methods

Stepwise discriminant analysis statistical method

For stepwise discriminant analysis (SDA), the specific operating approaches are as follows: All traits are selected as the input variables of the algorithm. Then, the SDA algorithm will select a variable that has the most significant discriminant ability. Next, the selecting for second variable based on the first variable, which indicates that combining the first and second variables will have the most significant discriminant ability. By that analogy, the third variable will be selected. Because of the mutual relationship between different variables, the previous variable may lose significant discriminant ability after inputting the new variable. Then, we will inspect the discriminant ability of all previous selected variables to find the disabled variables, remove them, and go on to find new variables until no significant variables can be removed. In this study, stepwise discriminant analysis training was achieved using (SPSS v.22 software), which is a proven technique for meaningfully classifying different shapes [69]. The detailed description is shown in Additional file 14.

Support vector machine statistical method

The support vector machine (SVM) is a common supervised learning algorithm that has been shown to provide state-of-the-art performance in many classification problems. The main thinking of the SVM is to establish a classification hyperplane as the decision curved surface, which maximizes the gap of positive samples and negative samples. The LIBSVM-matlab toolkit [70] was used here to conduct SVM model. All the 801 samples with three different leaf shapes were randomly divided into two groups (402 samples for training and 399 samples for testing). Firstly, we should limit all data into a certain range. Here, the interval from 0 to 1. The purpose for data normalization [71] is to ensure the convergence of the SVM algorithm. At the same time, it will improve the accuracy of classification. Next, we can train the model. Here, the kernel function is generated to use polynomials and the kernel parameter is set to 1.5. Also, the penalty parameter is set to 2. The genetic algorithm (GA) was used to choose the best value of kernel parameter and penalty parameter. The detailed algorithm process refers to Additional file 14 and the code is shown in Additional file 15.

Random forest statistical method

The random forest (RF) classifier is a combination of multiple decision trees. In this study, the open source randomforest-matlab toolkit [72] was adopted to build random forest classifier. Abhishek Jaiantilal, of the University of Colorado, Boulder, is the primary developer. Here, the number of decision trees in my random forest is 1000 and the other parameters adopt the default value. 801 samples were randomly selected and divided into two groups: 402 samples comprising the training group and 399 samples for testing group. When the test samples enter into the random forest, every decision tree will independently classify the category it belongs to. The final statistical classification result will abide by the majority rule. The detailed algorithm process refers to Additional file 14 and the code is shown in Additional file 16.

Authors’ contributions

XX developed the image analysis pipeline, performed the evaluation experiment, analyzed the data, and drafted the manuscript. LY and WY designed and built the hardware, analyzed the data, performed the evaluation experiment, and contributed in writing the manuscript. ML planted and managed the experimental material. NJ contributed in developed image analysis pipeline. DW provided support with hardware development and implementation of the evaluation experiment. GC planted and managed the experimental material and guided experiment. LX contributed in writing the manuscript. KL contributed in writing the manuscript. QL supervised the study and contributed in writing the manuscript. All authors read and approved the final manuscript.

Acknowledgements

This work was supported by Grants from the National Program on High Technology Development (2013AA102403), and the Scientific Conditions and Resources Research Program of Hubei Province of China (2015BCE044). We would like to thank Dr. Jiang for technical assistance.

Competing interests

The authors declare that they have no competing interests.

Availability of data and materials

All data generated during this study are included in this published article.

Abbreviations

AA

the aspect ratio of leaf minimum circumscribed box

R

the ratio of leaf area to minimum circumscribed box area

AC

the ratio of leaf area to leaf convex hull area

PC

the ratio of leaf circumference to leaf convex hull perimeter

S

the ratio of leaf area to the square of leaf convex hull perimeter

E

the ratio of long axis of ellipse to short axis of ellipse

FF

the ratio of leaf area to the square of leaf perimeter

PAR

the ratio of leaf circumference to leaf area

SFD

reflect the effectiveness of the occupies space without image cropping

IFD

reflect the effectiveness of the occupies space with image cropping

C

the ratio of inscribed circle radius to circumscribed circle radius

NIC

the number of inner cavities

APIC

the average perimeter of inner cavities

AAIC

the average area of inner cavities

TNI

total number of indents

ENI

effective number of indents

ADI

the average depth of indents

ADEI

the average depth of effective indents

AVE

the average calculated value of effectiveness

SDA

stepwise discriminant analysis

SVM

support vector machine 

RF

random forest

GA

genetic algorithm

Additional files

13007_2017_157_MOESM1_ESM.tif (9.5MB, tif)

Additional file 1: Figure S3. The software interface for seedling rape canopy three-dimensional reconstruction.

13007_2017_157_MOESM2_ESM.xls (15.2KB, xls)

Additional file 2. Original data of leaf area and canopy height.

13007_2017_157_MOESM3_ESM.xls (162.7KB, xls)

Additional file 3. Original data of individual leaf traits.

13007_2017_157_MOESM4_ESM.tif (7.3MB, tif)

Additional file 4: Figure S4. The software interface for extracting individual leaf traits.

Download video file (9.2MB, mp4)

Additional file 5. Supplementary document for manual interaction.

13007_2017_157_MOESM6_ESM.doc (1.7MB, doc)

Additional file 6. Supplementary document for overlap recovery algorithm.

13007_2017_157_MOESM7_ESM.tif (4.9MB, tif)

Additional file 7: Figure S5. Overlap recovery algorithm flow.

13007_2017_157_MOESM8_ESM.tif (9.3MB, tif)

Additional file 8: Figure S2. Three different shapes of rape leaf in seedling stage.

13007_2017_157_MOESM9_ESM.tif (3.8MB, tif)

Additional file 9: Figure S6. The occluded situation in binocular stereo-imaging system.

13007_2017_157_MOESM10_ESM.doc (88.5KB, doc)

Additional file 10. Supplementary document for three-dimensional reconstruction.

13007_2017_157_MOESM11_ESM.tif (6.7MB, tif)

Additional file 11: Figure S1. The triangle range finding for optical path.

13007_2017_157_MOESM12_ESM.txt (1.3KB, txt)

Additional file 12. Supplementary document for Delaunay code.

13007_2017_157_MOESM13_ESM.doc (75.5KB, doc)

Additional file 13. Supplementary document for individual leaf traits.

13007_2017_157_MOESM14_ESM.doc (270KB, doc)

Additional file 14. Supplementary document for three classification methods.

13007_2017_157_MOESM15_ESM.txt (3KB, txt)

Additional file 15. Supplementary document for SVM code.

13007_2017_157_MOESM16_ESM.txt (11.2KB, txt)

Additional file 16. Supplementary document for RF code.

Footnotes

Xiong Xiong and Lejun Yu contributed equally to this work

Contributor Information

Xiong Xiong, Email: 971123916@qq.com.

Lejun Yu, Email: yu@mail.hust.edu.cn.

Wanneng Yang, Email: ywn@mail.hzau.edu.cn.

Meng Liu, Email: 978431219@qq.com.

Ni Jiang, Email: 403347116@qq.com.

Di Wu, Email: 380524590@qq.com.

Guoxing Chen, Email: chenguoxing@mail.hzau.edu.cn.

Lizhong Xiong, Email: lizhongx@mail.hzau.edu.cn.

Kede Liu, Email: kdliu@mail.hzau.edu.cn.

Qian Liu, Phone: (86)-27-8779-2033, Email: qianliu@mail.hust.edu.cn.

References

  • 1.Paulauskas A, Jodinskienė M, Griciuvienė L, et al. Morphological traits and genetic diversity of differently overwintered oilseed rape (Brassica napus L.) cultivars. Zemdirb Agric. 2013;100(4):409–416. doi: 10.13080/z-a.2013.100.052. [DOI] [Google Scholar]
  • 2.Lorin M, Jeuffroy MH, Butier A, et al. Undersowing winter oilseed rape with frost-sensitive legume living mulch: Consequences for cash crop nitrogen nutrition. Field Crops Res. 2016;193:24–33. doi: 10.1016/j.fcr.2016.03.002. [DOI] [Google Scholar]
  • 3.Pospišil M, Brčic M, Husnjak S. Suitability of soil and climate for oilseed rape production in the Republic of Croatia. Agric Conspec Sci. 2011;76(1):35–39. [Google Scholar]
  • 4.Pessel D, Lecomte J, Emeriau V, et al. Persistence of oilseed rape (Brassica napus L.) outside of cultivated fields. Theor Appl Genet. 2001;102(6–7):841–846. doi: 10.1007/s001220100583. [DOI] [Google Scholar]
  • 5.Rathke GW, Diepenbrock W. Energy balance of winter oilseed rape (Brassica napus L.) cropping as related to nitrogen supply and preceding crop. Eur J Agron. 2006;24(1):35–44. doi: 10.1016/j.eja.2005.04.003. [DOI] [Google Scholar]
  • 6.Powers S, Pirie E, Latunde-Dada A, et al. Analysis of leaf appearance, leaf death and phoma leaf spot, caused by Leptosphaeria maculans, on oilseed rape (Brassica napus) cultivars. Ann Appl Biol. 2010;157(1):55–70. doi: 10.1111/j.1744-7348.2010.00409.x. [DOI] [Google Scholar]
  • 7.Bylesjö M, Segura V, Soolanayakanahally RY, Rae AM, Trygg J, Gustafsson P, Jansson S, Street NR. LAMINA: a tool for rapid quantification of leaf size and shape parameters. BMC Plant Biol. 2008;8(1):82. doi: 10.1186/1471-2229-8-82. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Efroni I, Eshed Y, Lifschitz E. Morphogenesis of simple and compound leaves: a critical review. Plant Cell. 2010;22(4):1019–1032. doi: 10.1105/tpc.109.073601. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Yang W, Duan L, Chen G, et al. Plant phenomics and high-throughput phenotyping: accelerating rice functional genomics using multidisciplinary technologies. Curr Opin Plant Biol. 2013;16(2):180–187. doi: 10.1016/j.pbi.2013.03.005. [DOI] [PubMed] [Google Scholar]
  • 10.Pérez-Pérez JM, Esteve-Bruna D, Micol JL. QTL analysis of leaf architecture. J Plant Res. 2010;123(1):15–23. doi: 10.1007/s10265-009-0267-z. [DOI] [PubMed] [Google Scholar]
  • 11.Cope JS, Corney D, Clark JY, et al. Plant species identification using digital morphometrics: a review. Expert Syst Appl. 2012;39(8):7562–7573. doi: 10.1016/j.eswa.2012.01.073. [DOI] [Google Scholar]
  • 12.Muller-Linow M, Pinto-Espinosa F, Scharr H, et al. The leaf angle distribution of natural plant populations: assessing the canopy with a novel software tool. Plant Methods. 2015;11:11. doi: 10.1186/s13007-015-0052-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Chéné Y, Rousseau D, Lucidarme P, et al. On the use of depth camera for 3D phenotyping of entire plants. Comput Electron Agric. 2012;82:122–127. doi: 10.1016/j.compag.2011.12.007. [DOI] [Google Scholar]
  • 14.Omasa K, Hosoi F, Konishi A. 3D lidar imaging for detecting and understanding plant responses and canopy structure. J Exp Bot. 2007;58(4):881–898. doi: 10.1093/jxb/erl142. [DOI] [PubMed] [Google Scholar]
  • 15.Shibayama M, Sakamoto T, Takada E, et al. Estimating paddy rice leaf area index with fixed point continuous observation of near infrared reflectance using a calibrated digital camera. Plant Prod Sci. 2011;14(1):30–46. doi: 10.1626/pps.14.30. [DOI] [Google Scholar]
  • 16.Tsukaya H. Organ shape and size: a lesson from studies of leaf morphogenesis. Curr Opin Plant Biol. 2003;6(1):57–62. doi: 10.1016/S1369526602000055. [DOI] [PubMed] [Google Scholar]
  • 17.Neto JC, Meyer GE, Jones DD. Individual leaf extractions from young canopy images using Gustafson–Kessel clustering and a genetic algorithm. Comput Electron Agric. 2006;51(1–2):66–85. doi: 10.1016/j.compag.2005.11.002. [DOI] [Google Scholar]
  • 18.Bama BS, Valli SM, Raju S, et al. Content based leaf image retrieval (CBLIR) using shape, color and texture features. Indian J Comput Sci Eng. 2011;2(2):202–211. [Google Scholar]
  • 19.Nam Y, Hwang E, Kim D. A similarity-based leaf image retrieval scheme: joining shape and venation features. Comput Vis Image Underst. 2008;110(2):245–259. doi: 10.1016/j.cviu.2007.08.002. [DOI] [Google Scholar]
  • 20.O’Neal ME, Landis DA, Isaacs R. An inexpensive, accurate method for measuring leaf area and defoliation through digital image analysis. J Econ Entomol. 2002;95(6):1190–1194. doi: 10.1603/0022-0493-95.6.1190. [DOI] [PubMed] [Google Scholar]
  • 21.Bakr EM. A new software for measuring leaf area, and area damaged by Tetranychus urticae Koch. J Appl Entomol. 2005;129(3):173–175. doi: 10.1111/j.1439-0418.2005.00948.x. [DOI] [Google Scholar]
  • 22.Igathinathane C, Prakash VSS, Padma U, et al. Interactive computer software development for leaf area measurement. Comput Electron Agric. 2006;51(1–2):1–16. doi: 10.1016/j.compag.2005.10.003. [DOI] [Google Scholar]
  • 23.Weight C, Parnham D, Waites R. Technical advance: leafAnalyser: a computational method for rapid and large-scale analyses of leaf shape variation. Plant J. 2008;53(3):578–586. doi: 10.1111/j.1365-313X.2007.03330.x. [DOI] [PubMed] [Google Scholar]
  • 24.Dengkui A, Minzan L, Li Z. Measurement of tomato leaf area using computer image processing technology. Sensor Lett. 2010;8(1):56–60. doi: 10.1166/sl.2010.1201. [DOI] [Google Scholar]
  • 25.Yang W, Guo Z, Huang C, et al. Genome-wide association study of rice (Oryza sativa L.) leaf traits with a high-throughput leaf scorer. J Exp Bot. 2015;66(18):5605–5615. doi: 10.1093/jxb/erv100. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26.Paulus S, Behmann J, Mahlein AK, et al. Low-cost 3D systems: suitable tools for plant phenotyping. Sensors (Basel) 2014;14(2):3001–3018. doi: 10.3390/s140203001. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Frasson RPM, Krajewski WF. Three-dimensional digital model of a maize plant. Agric For Meteorol. 2010;150(3):478–488. doi: 10.1016/j.agrformet.2010.01.003. [DOI] [Google Scholar]
  • 28.Dornbusch T, Lorrain S, Kuznetsov D, et al. Measuring the diurnal pattern of leaf hyponasty and growth in Arabidopsis—a novel phenotyping approach using laser scanning. Funct Plant Biol. 2012;39(11):860. doi: 10.1071/FP12018. [DOI] [PubMed] [Google Scholar]
  • 29.Chen S, Li YF, Zhang J, et al. Active sensor planning for multiview vision tasks. Berlin: Springer; 2008. [Google Scholar]
  • 30.Henry P, Krainin M, Herbst E, et al. RGB-D mapping: using Kinect-style depth cameras for dense 3D modeling of indoor environments. Int J Robot Res. 2012;31(5):647–663. doi: 10.1177/0278364911434148. [DOI] [Google Scholar]
  • 31.Baumberg A, Lyons A, Taylor R. 3D S.O.M.—a commercial software solution to 3D scanning. Graph Models. 2005;67(6):476–495. doi: 10.1016/j.gmod.2004.10.002. [DOI] [Google Scholar]
  • 32.Zhou Z. 3D reconstruction from multiple images using single moving camera [D]. 2015.
  • 33.Park J-S. Interactive 3D reconstruction from multiple images: a primitive-based approach. Pattern Recogn Lett. 2005;26(16):2558–2571. doi: 10.1016/j.patrec.2005.05.009. [DOI] [Google Scholar]
  • 34.Herman M, Kanade T. Incremental reconstruction of 3D scenes from multiple, complex images. Artif Intell. 1986;30(3):289–341. doi: 10.1016/0004-3702(86)90002-0. [DOI] [Google Scholar]
  • 35.Seitz S M, Curless B, Diebel J, et al. A comparison and evaluation of multi-view stereo reconstruction algorithms. In: IEEE computer society conference on computer vision and pattern recognition; 2006. p. 519–528.
  • 36.Scharstein D, Szeliski R. A taxonomy and evaluation of dense two-frame stereo correspondence algorithms: stereo and multi-baseline vision. New York: IEEE; 2002. pp. 131–140. [Google Scholar]
  • 37.Kise M, Zhang Q. Creating a panoramic field image using multi-spectral stereovision system. Comput Electron Agric. 2008;60(1):67–75. doi: 10.1016/j.compag.2007.07.002. [DOI] [Google Scholar]
  • 38.Xia C, Li Y, Chon TS et al. A stereo vision based method for autonomous spray of pesticides to plant leaves. In: IEEE international symposium on industrial electronics. IEEE; 2009. p. 909–914.
  • 39.Xiang R, Jiang H, Ying Y. Recognition of clustered tomatoes based on binocular stereo vision. Comput Electron Agric. 2014;106:75–90. doi: 10.1016/j.compag.2014.05.006. [DOI] [Google Scholar]
  • 40.Wang C, Zou X, Tang Y, et al. Localisation of litchi in an unstructured environment using binocular stereo vision. Biosyst Eng. 2016;145:39–51. doi: 10.1016/j.biosystemseng.2016.02.004. [DOI] [Google Scholar]
  • 41.Ivanov N, Boissard P, Chapron M, et al. Computer stereo plotting for 3-D reconstruction of a maize canopy. Agric For Meteorol. 1995;75(1–3):85–102. doi: 10.1016/0168-1923(94)02204-W. [DOI] [Google Scholar]
  • 42.Andersen HJ, Reng L, Kirk K. Geometric plant properties by relaxed stereo vision using simulated annealing. Comput Electron Agric. 2005;49(2):219–232. doi: 10.1016/j.compag.2005.02.015. [DOI] [Google Scholar]
  • 43.Biskup B, Scharr H, Schurr U, et al. A stereo imaging system for measuring structural parameters of plant canopies. Plant Cell Environ. 2007;30(10):1299–1308. doi: 10.1111/j.1365-3040.2007.01702.x. [DOI] [PubMed] [Google Scholar]
  • 44.Biskup B, Scharr H, Fischbach A, et al. Diel growth cycle of isolated leaf discs analyzed with a novel, high-throughput three-dimensional imaging method is identical to that of intact leaves. Plant Physiol. 2009;149(3):1452–1461. doi: 10.1104/pp.108.134486. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 45.Lou L, Liu Y, Han J, et al. Accurate multi-view stereo 3D reconstruction for cost-effective plant phenotyping. In: International Conference, Iciar; 2014. p. 349–356.
  • 46.Rose JC, Paulus S, Kuhlmann H. Accuracy analysis of a multi-view stereo approach for phenotyping of tomato plants at the organ level. Sensors. 2015;15(5):9651–9665. doi: 10.3390/s150509651. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 47.Miller J, Morgenroth J, Gomez C. 3D modelling of individual trees using a handheld camera: accuracy of height, diameter and volume estimates. Urban For Urban Green. 2015;14(4):932–940. doi: 10.1016/j.ufug.2015.09.001. [DOI] [Google Scholar]
  • 48.Yang W, Guo Z, Huang C, et al. Combining high-throughput phenotyping and genome-wide association studies to reveal natural genetic variation in rice. Nat Commun. 2014;5:5087. doi: 10.1038/ncomms6087. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 49.Song Y, Wilson R, Edmondson R, et al. Surface modelling of plants from stereo images. 2007;312–319:312–319. [Google Scholar]
  • 50.Edelsbrunner H, Seidel R. Voronoi diagrams and arrangements. Discr Comput Geom. 1986;1(1):25–44. doi: 10.1007/BF02187681. [DOI] [Google Scholar]
  • 51.Farooq M, Tagle AG, Santos RE, et al. Quantitative trait loci mapping for leaf length and leaf width in rice cv. IR64 derived lines. J Integr Plant Biol. 2010;52(6):578–584. doi: 10.1111/j.1744-7909.2010.00955.x. [DOI] [PubMed] [Google Scholar]
  • 52.Duan L, Yang W, Huang C, et al. A novel machine-vision-based facility for the automatic evaluation of yield-related traits in rice. Plant Methods. 2011;7(1):1–13. doi: 10.1186/1746-4811-7-44. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 53.Duan L, Huang C, Chen G, et al. High-throughput estimation of yield for individual rice plant using multi-angle RGB imaging. In: Computer and computing technologies in agriculture VIII. Springer, Berlin; 2015. p. 1–12.
  • 54.Iosifidis A, Gabbouj M. Multi-class support vector machine classifiers using intrinsic and penalty graphs. Pattern Recogn. 2016;55:231–246. doi: 10.1016/j.patcog.2016.02.002. [DOI] [Google Scholar]
  • 55.Shi Z, Cheng J-L, Huang M-X, et al. Assessing reclamation levels of coastal saline lands with integrated stepwise discriminant analysis and laboratory hyperspectral data. Pedosphere. 2006;16(2):154–160. doi: 10.1016/S1002-0160(06)60038-6. [DOI] [Google Scholar]
  • 56.Fung GM, Mangasarian OL. Multicategory proximal support vector machine classifiers. Mach Learn. 2005;59(1–2):77–97. doi: 10.1007/s10994-005-0463-6. [DOI] [Google Scholar]
  • 57.Breiman L. Random forests. Mach Learn. 2001;45(1):5–32. doi: 10.1023/A:1010933404324. [DOI] [Google Scholar]
  • 58.Pardo M, Sberveglieri G. Random forests and nearest shrunken centroids for the classification of sensor array data. Sensors Actuat B Chem. 2008;131(1):93–99. doi: 10.1016/j.snb.2007.12.015. [DOI] [Google Scholar]
  • 59.Bai X, Sun C, Zhou F. Splitting touching cells based on concave points and ellipse fitting. Pattern Recogn. 2009;42(11):2434–2446. doi: 10.1016/j.patcog.2009.04.003. [DOI] [Google Scholar]
  • 60.Awrangjeb M, Lu G, Fraser C S, et al. A fast corner detector based on the chord-to-point distance accumulation technique. In: Digital image computing: techniques and applications. IEEE Computer Society; 2009. p. 519–525
  • 61.White NDG, Jayas DS, Gong Z. Separation of touching grain kernels in an image by ellipse fitting algorithm. Biosyst Eng. 2005;92(2):135–142. doi: 10.1016/j.biosystemseng.2005.06.010. [DOI] [Google Scholar]
  • 62.Jean-Yves B. Camera calibration toolbox for matlab [EB/OL]. http://www.vision.caltech.edu/. 12, 2011.
  • 63.Jiang N, Yang W, Duan L, et al. A nondestructive method for estimating the total green leaf area of individual rice plants using multi-angle color images. J Innov Opt Health Sci. 2015;08(02):1550002. doi: 10.1142/S1793545815500029. [DOI] [Google Scholar]
  • 64.Geiger A, Roser M, Urtasun R. Efficient large-scale stereo matching. In: Computer vision—ACCV 2010. Springer, Berlin; 2010. p. 25–38.
  • 65.Cochran SD, Medioni G. 3-D surface description from binocular stereo. IEEE Trans Pattern Anal Mach Intell. 1992;14(10):981–994. doi: 10.1109/34.159902. [DOI] [Google Scholar]
  • 66.Fua P. A parallel stereo algorithm that produces dense depth maps and preserves image features. Mach Vis Appl. 1993;6(1):35–49. doi: 10.1007/BF01212430. [DOI] [Google Scholar]
  • 67.Woebbecke D, Meyer G, Von Bargen K, et al. Color indices for weed identification under various soil, residue, and lighting conditions. Trans ASAE. 1995;38(1):259–269. doi: 10.13031/2013.27838. [DOI] [Google Scholar]
  • 68.Meyer G, Mehta T, Kocher M, et al. Textural imaging and discriminant analysis for distinguishingweeds for spot spraying. Trans ASAE. 1998;41(4):1189. doi: 10.13031/2013.17244. [DOI] [Google Scholar]
  • 69.Petalas C, Anagnostopoulos K. Application of Stepwise discriminant analysis for the identification of salinity sources of groundwater. Water Resour Manage. 2006;20(5):681–700. doi: 10.1007/s11269-005-9002-2. [DOI] [Google Scholar]
  • 70.Chang C-C, Lin C-J. LIBSVM: a library for support vector machines. ACM Trans Intell Syst Technol (TIST) 2011;2(3):27. [Google Scholar]
  • 71.Navarro PJ, Fernando P, Julia W, et al. Machine learning and computer vision system for phenotype data acquisition and analysis in plants. Sensors. 2016;16(5):641. doi: 10.3390/s16050641. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 72.Liaw A, Wiener M. Classification and regression by random forest. R News 2001; 23(23).

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

All data generated during this study are included in this published article.


Articles from Plant Methods are provided here courtesy of BMC

RESOURCES