Skip to main content
Frontiers in Plant Science logoLink to Frontiers in Plant Science
. 2026 Apr 10;17:1790984. doi: 10.3389/fpls.2026.1790984

Cotton field population phenotyping analysis based on 3D Gaussian reconstruction and dynamic spatial constraints

Bo Liu 1, Xiaojuan Li 1,*, Jiajie Yang 2, Zhi Liang 1, Xiaotong Zheng 1, Xiangjun Zou 1,*
PMCID: PMC13106366  PMID: 42040288

Abstract

Introduction

High-throughput field phenotyping (HTFP) holds great potential for elucidating the relationship between genomes and phenotypes. However, obtaining high-quality three-dimensional point cloud data of field populations and achieving single-plant phenotypic analysis remain challenging.

Methods

This study develops an integrated framework for field crop reconstruction based on 3D Gaussian splatting, incorporating a geometry-aware dynamic constraint algorithm to achieve instance segmentation and extract key phenotypic traits of individual plants. Using 3D Gaussian splatting technology, field-scale cotton population modeling is accomplished, generating dense 3D point clouds for regions of interest. Furthermore, the concept of a crop localization domain is proposed, establishing a longitudinal mapping that associates plant positional coordinates with long-term phenotypic attributes. Finally, through a dynamic spatial constraint mechanism, the accuracy and computational efficiency of instance segmentation for crop population point clouds are significantly improved, enabling rapid extraction of individual plant traits such as cotyledon node height, plant height, and leaf area.

Results

The results demonstrate that PhenotypeAI successfully reconstructed nine cotton populations with PSNR exceeding 30.0 dB. It successfully extracted regions of interest from 403 cotton plants, achieving an average F-score of 91.32% for instance segmentation and an average accuracy of 91.35%. The extracted traits—cotyledon node height, plant height, and leaf area—exhibited strong correlations with manual measurements, with coefficients of determination (R2) of 0.90, 0.91, and 0.91, respectively.

Discussion

The proposed method provides a low-cost solution for high-throughput field phenotypic analysis of field cotton and improves the efficiency of cotton breeding.

Keywords: cotton, deep learning, plant phenotyping, point cloud segmentation, semantic segmentation, three-dimensional reconstruction

1. Introduction

Cotton, as one of the world’s most important agricultural products, is the most widely cultivated natural fiber crop globally (Jiang et al., 2025; Yu et al., 2025). It finds extensive applications in textiles, food, medicine, construction, and other sectors, holding significant importance in international trade (Li et al., 2020). Faced with challenges such as climate change and resource depletion, there is an urgent need to develop high-quality cotton varieties to promote the high-quality development of the cotton industry (Abdurakhimov et al., 2023). Cotton breeding technology requires the evaluation of multiple traits across thousands of breeding lines through multi-year, multi-location trials with multiple replicates, screening high-quality varieties from vast sample populations (Chawade et al., 2019; Dipta et al., 2023). Consequently, the rapid and accurate acquisition and analysis of plant phenotypic information constitute a critical component of variety selection and breeding (He et al., 2024). However, cotton exhibits a vast array of varieties, rich phenotypic information, and complex biological characteristics (Chen et al., 2024; Iqbal et al., 2020). The acquisition of its phenotypic information has long remained confined to traditional manual methods, characterized by low throughput, inconsistent standards, and typically destructive, labor-intensive, and time-consuming procedures (Bian et al., 2022), making it challenging to accurately obtain phenotypic data for individual crops at the population scale (Fu and Jiang, 2022). Furthermore, manual measurement of phenotypic parameters typically allows sampling only at fixed time points, failing to provide continuous monitoring of the dynamic developmental processes and variation characteristics within the same plant. There is an urgent need for automated, high-precision, and low-cost technologies for phenotyping acquisition and analysis in field trial plots.

Visual-based target and crop identification and detection methods, primarily encompassing deep learning (Tang et al., 2020), point cloud reconstruction (Li et al., 2024; Peng et al., 2025), and color segmentation techniques (Liang et al., 2025), have achieved significant progress. Notably, the integration of 3D point cloud reconstruction with plant analysis technology is emerging as an automated alternative for capturing phenotypic information. While LiDAR scanning (Chen et al., 2025; Gu et al., 2024; Ma et al., 2024) can precisely capture crop phenotypic parameters, its low efficiency, high cost, and lack of fine-grained details in reconstructed plant point clouds limit its applicability for phenotypic measurements. Structure from Motion (SFM) and Multi-View Stereo (MVS) estimate scene structure by triangulating keypoints across multiple images. These methods have been widely applied in extracting phenotypic parameters for crops such as soybean (Cui et al., 2025; He et al., 2023), sugar beet (Chen et al., 2024), and maize (Gao et al., 2023). However, due to limitations in feature matching and dense reconstruction algorithms, this approach often fails to generate complete and detailed point clouds. Neural Radiation Fields (NeRF) (Tancik et al., 2023) and other implicit neural representation techniques demonstrate significant potential in overcoming the limitations of traditional Structure from Motion-Model-Based Visualization (SfM-MVS) methods. They can reconstruct high-quality 3D models with superior performance in occluded regions and plant detail. When applied to the three-dimensional reconstruction of field populations, conventional methods based on Structure-from-Motion (SFM) and Multi-View Stereo (MVS) using RGB imagery tend to generate considerable noise along leaf margins and produce fractured point clouds for plant stems (Gu et al., 2024). NeRF-based methods effectively reduce edge noise and mitigate point cloud fragmentation by leveraging neural networks to learn scene geometry and appearance. However, this approach requires intensive sampling of the neural network to generate point clouds, making it more suitable for small-scale studies or simple trait measurements on limited plant populations (F. Saeed et al., 2023). Three-dimensional Gaussian Splatter (3DGS) (Bao et al., 2025; Kerbl et al., 2023), a point-based 3D representation technique, constructs spatially distributed Gaussian primitives based on observed viewpoints. This generates explicit expressions of scene color distribution and texture features, enabling real-time rendering while maintaining high visual quality. It has emerged as a more efficient 3D reconstruction solution. Currently, with breakthrough advancements in algorithms such as Neural Radiance Fields (NeRF) and 3D Gaussian Splatting, multi-view imaging using smartphones has demonstrated significant potential for 3D reconstruction at the organ level in plants. This enables high-quality 3D reconstruction of plant populations in field conditions, offering substantial advantages in terms of low cost and convenience.

Obtaining single plant crops from field populations through point cloud instance segmentation is a crucial prerequisite for automatic, accurate, and non-invasive acquisition of plant phenotypic information (Digumarti et al., 2018; Elnashef et al., 2019; Wang et al., 2025a). Currently, the main methods for instance segmentation of point clouds in crop populations include those based on classical algorithms (Mirande et al., 2022; Zou et al., 2023) and those based on deep learning networks (Dong et al., 2025; J and Nidamanuri, 2024; Li et al., 2023; Wang et al., 2025b). Mirande et al. (2022) achieved plant point cloud segmentation and extracted target phenotypic parameters based on agronomic knowledge. Chen et al. (2025) employed the PointNet++ method to achieve semantic segmentation of LiDAR point clouds in cotton fields, successfully extracting cotton point clouds at both the single-plant and plot scales. Wang et al. (2025b) proposed the rubber tree point cloud segmentation network RsegNet, achieving a best F-score of 86.1% for rubber tree segmentation on the FOR-instance forest dataset. Deep learning has demonstrated impressive performance in plant point cloud instance segmentation, but its training relies on massive, high-precision annotated datasets. Zhang et al. (2025) proposed an innovative method named Wheat3DGS, which integrates 3D Gaussian splatting (3DGS) with the Segment Anything Model (SAM), to achieve field point cloud reconstruction of wheat canopy parts, instance segmentation of wheat spike organs, and phenotypic measurement. The method uses a pre-trained wheat spike detector (YOLOv5) to generate SAM prompts for obtaining 2D segmentation masks, projects these masks into the 3DGS-reconstructed scene, and achieves multi-view mask association and 3D instance segmentation through an iterative matching fine-tuning strategy. This method targets the wheat canopy spike parts, which are characterized by relatively uniform morphology, dense distribution, and compactness. In contrast, field cotton populations present challenges due to inter-plant branch entanglement, canopy adhesion, and significant variation in appearance features such as leaf orientation and branch morphology, lacking stable and consistent appearance characteristics. This determines that the strategy of transferring YOLO detectors and 2D mask projection, which relies on appearance priors, is difficult to apply effectively to cotton. Achieving precise semantic segmentation of complete cotton plants in high-density, compact cotton field environments remains a challenging task. Existing methods have not fully addressed the challenges frequently encountered when segmenting individual cotton plants, especially in high-density and compact cotton field environments. The main issues are as follows: (1) In high-density cotton populations, inter-plant branch entanglement, canopy adhesion, and variable leaf orientation and occlusion states result in a lack of stable, distinguishable appearance features. This complexity makes it difficult for appearance-based deep learning models to learn robust feature representations. More critically, performing pixel-level or instance-level manual precise annotation for complete field cotton population point clouds is difficult, costly, and error-prone, limiting the feasibility of supervised learning frameworks. (2) As a densely planted, low-growing crop, cotton plants within a population exhibit high similarity in three-dimensional geometric morphology, while also having uneven planting spacing and potential random offsets between plants. This makes traditional unsupervised methods that rely solely on local point cloud density or general geometric features highly prone to over-segmentation or under-segmentation at plant boundaries. (3) Despite macroscopic similarity among plants in the population, individual plants still exhibit differences in structural parameters such as height, canopy width, and branching angle. Segmentation methods that ignore the vertical distribution of crops and the topological properties of individual plant morphology struggle to accurately define plant boundaries, reducing segmentation accuracy and reliability. This paper focuses on the inherent, stable prior knowledge and topological properties of the geometric structure of cotton plants, and implements a training-free, annotation-free instance segmentation method.

In summary, this study develops an integrated framework for field crop reconstruction based on 3D Gaussian Splatting, incorporating a geometry-aware dynamic constraint algorithm to achieve instance segmentation. This framework can accurately capture the three-dimensional morphological structure of individual cotton plants, which is of significant importance for the precise measurement of cotton phenotypic parameters and breeding work. The main contributions are as follows:

  1. A high-quality 3D point cloud reconstruction pipeline for field cotton populations is proposed. Based on 3D Gaussian Splatting (3DGS) for cotton population reconstruction, point cloud data is sampled via Gaussian distribution probability, and outliers are eliminated using a Mahalanobis distance threshold, ultimately generating high-density, high-precision point clouds of cotton populations.

  2. We introduced the concept of a plant localization zone. By generating non-convex clusters from the dataset using the Density-Based Spatial Clustering of Applications with Noise (DBSCAN) algorithm, we determined the location of each cotton plant and established a longitudinal mapping relationship between plant position information and phenotypic data.

  3. We incorporated the vertical distribution and morphological structure of cotton plants within a group into the constraint scope through a dynamic spatial constraint approach. This enabled the establishment of an improved regional growth method for segmenting cotton plant groups, allowing the group point cloud to rapidly segment plant point clouds within a predetermined constrained space while adapting to variations in the size of the cotton group.

  4. To validate the method’s effectiveness, extensive experiments were conducted. Quantitative and qualitative analyses were performed on field-collected cotton group datasets to evaluate reconstruction and segmentation performance. Point cloud data from 403 multi-varietal cotton plants were successfully acquired. Based on segmentation results, key structural parameters of individual plants—including cotyledon node height, plant height, leaf count, and leaf area—were successfully measured.

2. Materials and methods

2.1. Field trial design and cotton plant material

The experiment was conducted at the Kuqa Modern Agricultural Breeding Center in China (longitude: 82° 97', latitude: 41° 68'). The breeding field implemented lean agronomic management, including regular chemical weeding and manual inspection, resulting in only sporadic and minor weeds in the field. Prior to planting, land leveling and soil preparation were carried out, and plastic film mulch was laid, making the field surface relatively flat. The planting pattern is shown in Figure 1, featuring plot cultivation of different varieties. Breeding experts selected three new varieties suitable for mechanical harvesting from the breeding field: ‘Kechuang 94’, ‘Fuzimian 6’, and ‘Xinluzao 36’. Each variety was arranged in three replicate plots, totaling nine plots. Each sample plot measured 2.0 m × 2.0 m. Precision seeding was performed using a precision seeder with uniform spacing: row spacing of 660 mm (25.98 inches) and plant spacing of 100 mm (3.94 inches), with one plant per hill. The effective plant counts were as follows: VARc94 (133 plants), VARf6 (132 plants), VARx36 (138 plants), with a total population size of 403 plants.

Figure 1.

Figure with three main panels detailing a cotton field data acquisition and analysis process. Panel a shows a map highlighting data acquisition locations in Xinjiang Uygur Autonomous Region and a cotton field layout with planting rows marked in red. Panel b displays a handheld gimbal with a smartphone used for video recording along three vertical paths at different heights—half, equal, and one and a half times the cotton population height. Panel c illustrates a workflow from a planting map through sample filtration, data mapping, quality audit, and decoding, resulting in organized cotton field image datasets.

Experimental fields and data acquisition. (a) Experiment location. (b) Sensor height and motion path during data acquisition. (c) Locate sample plots based on the planting map, acquire the sample plot envelope imagery data, and extract key frames.

The collected data pertains to the 6–7 leaf stage of cotton plants. Key phenotypic traits of focus during this period include plant height, cotyledon node height, number of leaves, and leaf area. Field data collection must be conducted under stable lighting conditions with wind speeds 3 m/s on the Beaufort scale. Avoid extreme weather and significant crop sway to ensure clear image textures and accurate point cloud registration. Operations should be conducted between 13:00 and 15:00. The experiment employs a stratified random sampling design: based on the spatial distribution map of the breeding field, nine independent sample plots are allocated to the three varieties VARc94, VARx36, and VARf6 in a checkerboard pattern. A minimum isolation zone of 2 m between plots is maintained to eliminate edge effects. The data acquisition system is built using an iPhone 12 smartphone terminal, with specific parameters detailed in Table 1. The operator moves clockwise along the sample area at a uniform pace, maintaining an average distance of approximately 0.8 meters between the acquisition terminal and the target cotton plants. Continuous 2K video recording is conducted for 120 seconds. Video data underwent frame parsing using FFmpeg 5.0, employing an equidistant sampling strategy to extract 840 frames per plot. Lens distortion correction and automatic white balance processing were performed via OpenCV 4.5, ultimately generating a 7,560 frame multi-perspective image dataset.

Table 1.

Key parameters of data acquisition platform.

Parameters Value
Sensor diagonal (mm) 7.62
Focal length (mm) 26
Image width (pixels) 1920
Image height (pixels) 1080
Angle of view (°) 120

Phenotypic Measurement Data Collection: The ST-500 stainless steel ruler was used to measure the plant height of cotton vertically from the cotyledon node of the main stem to the top growth point of the plant. The steel ruler was kept parallel to the axis of the main stem during the measurement. Each plant was measured three times repeatedly, and the arithmetic mean value was taken as the final value. Align the zero point of the tape measure with the stem base and follow the main stem’s curvature to measure the height to the first leaf node. Repeat this measurement three times, taking the arithmetic mean as the final value. Perform non-destructive leaf area measurements using the YMJ-D handheld live leaf area meter. Select 30 functional leaves per sample plot using systematic sampling, with 10 leaves each from the upper, middle, and lower canopy layers. To achieve precise matching of predicted leaves with manually measured samples, we utilized pre-established grid labels in field experiments (each experimental plot, row, and column has a unique identification number). During data collection, the specific insertion position at the X-th node from the top was recorded for each leaf used in manual measurements.

2.2. Image data preprocessing

Preprocessing of the collected raw video data primarily involves analyzing the pixel proportion of target cotton and the frame overlap rate between adjacent frames. Images with insufficient target area coverage and data segments with excessively high frame overlap rates are discarded. Camera pose estimation and sparse point cloud reconstruction were achieved using pose and position recovery technology. The principle is shown in Figure 2.

Figure 2.

Diagram illustrating a workflow for 3D plant modeling. Panel a shows how moving cameras capture feature points from different angles, panel b presents upper, middle, and lower plant images, and panel c visualizes image-derived 3D plant point clouds grouped by capture height.

Process of sparse point cloud construction: (a) Spatial point cloud prediction based on adjacent images through camera pose estimation, triangulation, and bundle adjustment; (b) Schematic diagram of image acquisition, divided into upper, middle, and lower acquisition angles; (c) Sparse point cloud obtained via motion recovery techniques.

2.3. Rapid 3D reconstruction technology for group crops

Three-dimensional reconstruction serves as a fundamental step in high-throughput phenotyping analysis of cotton fields, with modeling results directly impacting the reliability of trait quantification. Based on multi-view images captured, crop 3D reconstruction methods primarily include motion recovery techniques for multi-view stereo (MVS) (Pepe et al., 2022) and neural radiant fields (NeRF) (Mildenhall et al., 2021). Meanwhile, 3D Gaussian Splatter(3DGS) (Bao et al., 2025), an emerging scene rendering technique, has seen limited research in crop population modeling. This study evaluates the effectiveness of MVS, NeRF, and PhenotypeAI in crop population modeling using collected cotton canopy envelope data.

The technical implementation of 3D Gaussian modeling in cotton population reconstruction is essentially based on a synergistic mechanism of the four elements: geometric representation, differentiable rendering engine, adaptive optimization and dynamic topology regulation. The principle is shown in Figure 3. The mathematical basis stems from parameterizing the plant morphology as a collection of anisotropic Gaussian spheres (Equation 1). Each Gaussian cell defines its spatial distribution characteristics through the center position µ, the covariance matrix Σ, where Σ is parameterized by the rotation matrix R and the scaling matrix S to capture the organ-level growth characteristics of cotton, such as leaf blade stretching along the leaf vein direction, and stem geometry characteristics.

Figure 3.

Two-panel process flow diagram for plant 3D scene reconstruction using Gaussian representations. Panel (a) shows steps from plant sparse point initialization to Gaussian splatting, projection, adaptive density control, differentiable rasterization, and image comparison, with operation and gradient flows indicated. Panel (b) depicts generating a Gaussian rendering scene model and dense point cloud from field images, followed by rendering colors, calculating contributions, updating Gaussians, generating and validating points, removing outliers, and generating surface points with normal calculations. Insets display zoomed regions of plant imagery for clarity.

Three-dimensional Gaussian splatter and dense point cloud construction: (a) Initial Gaussian splatter is constructed, with adjustments to position, color, and transparency to achieve cotton cluster scene rendering; (b) Dense point cloud generation primarily involves Gaussian initialization and filtering, color rendering, point sampling, and outlier handling processes.

G(x)=exp (12(xμ)TΣ1(xμ)) (1)

where μ and Σ are the Gaussian center position and the covariance matrix, respectively, the latter of which mainly serves to control the shape and direction of the Gaussian distribution. The covariance matrix is decomposed by the rotation matrix R and the scaling matrix S as Equation 2:

Σ=RSSTRT (2)

where S denotes each axial scaling factor and R is the rotation matrix derived from the quaternion q. Based on this, the differentiable Gaussian sputtering technique projects the 3D Gaussian to the 2D image plane and establishes a bridge from parameter space to pixel space. The process achieves 3D to 2D geometric mapping through the camera projection matrix K with the view transformation W and calculates the color contribution of each pixel using the radial basis function property of the Gaussian kernel. And the projected covariance matrix Σ' is computed from the camera internal reference matrix K and the view transformation W with the core equation of Equation 3:

Σ=(JWΣWT)JT (3)

where J is the Jacobi matrix of the perspective projection. The output of the above differentiable pipeline drives the convergence process of the optimization objective function. The objective function consists of an L1 luminance loss, which is a direct measure of pixel-level color differences, weighted by a structural similarity loss (SSIM), and a structural similarity loss (SSIM), which enhances the sensitivity of the optimization to the details of the edges of the plant by comparing the luminance, contrast, and structural information within an image block.

2.4. Instance segmentation method for plant point cloud ensembles

This study proposes an instance segmentation strategy for cotton field point clouds, which primarily encompasses three integral components: preprocessing of cotton field scene point clouds, extraction of core points within plant localization zones, and spatially constrained segmentation of individual plant point clouds.

2.4.1. Step one: preprocessing of cotton field scene point clouds

To extract crucial regions of interest from a large volume of extraneous background point cloud data, as depicted by the red box in Figures 4a, a coordinate transformation is initially applied to the point cloud within the region of interest. Perform coordinate transformation on the point cloud of the region of interest, adjusting the ground-level point cloud to a horizontal orientation to achieve coordinate correction for cotton’s upward growth.

Figure 4.

Four-panel scientific diagram illustrates the workflow of processing a cotton population point cloud model, showing initial data, region of interest selection, noise filtering, and preprocessing steps, with inset close-ups and point counts annotated for each stage.

Preprocessing of cotton field scene point clouds. (a) Dense point cloud model of the cotton field constructed from the collective 3D reconstruction of VARf6 variety sample plot ‘a’. (b) Acquisition of the point cloud within the region of interest. (c) The region of interest point cloud after the application of statistical and bilateral filtering. (d) Implementation of an enhanced filtering methodology to facilitate the partial segregation of ground points (GS) from plant points (GP), followed by subsequent refinement processing of the GP component.

Based on the RANSAC algorithm, ground points are detected and a plane is fitted, yielding a plane equation ax + by + cz + d = 0. In this equation, (a, b, c) represents the normal vector of the plane, and (x, y, z) denotes the coordinates of any point on the plane. The normal vector of the 242 ground plane is computed and its direction is corrected. This ground normal vector n is then rotated to align with a target vector t, and the corresponding rotation quaternion q is constructed as Equation 4:

q=(w,x,y,z)=(cos θ2,sin θ2·kx,sin θ2·ky,sin θ2·kz) (4)

where, k=n×tn×t is the axis of rotation, and θ=arccos (n·t) is the angle of rotation. The rotation matrix for global coordinate correction R is then calculated using the quaternion:

R=[12y22z22xy2zw2xz+2yw2xy+2zw12x22z22yz2xw2xz2yw2yz+2xw12x22y2] (5)

Applying this rotation matrix to the entire point cloud within the region of interest completes the coordinate transformation for the cotton canopy. Next, the Statistical Outlier Removal algorithm and bilateral filtering method were employed to eliminate outlier points within the neighborhood of plant point clouds and other non-plant points (i.e., noise signals) from the region of interest (Figures 4b, c). Finally, a ground filter was developed based on the Cloth Simulation Filter (CSF) method, which segmented the cotton point cloud within the region of interest into ground plants (GP) and ground surface (GS). While originally designed for monitoring and surveying extremely large ground structures, we discovered that converting the cotton point cloud color to the HSI color space revealed a significant distinction between cotton and the ground in the H component. Therefore, the filter was optimized by augmenting the point cloud’s H component and increasing its nodes, resulting in the HCSF (HSI-enhanced Cloth Simulation Filter). The filtered cotton plant point cloud is shown in Figure 4d.

2.4.2. Step 2: extraction of core points for plant localization

Cotton canopy leaves tend to intersect as they compete for light resources, but the main stems naturally remain distinct. Therefore, specific regions of the cotton main stems can serve as localization areas for individual plants, enabling the calculation of a localization point for each plant (Figure 5). Unlike K-means, DBSCAN generates non-convex clusters based on the dataset and does not require the number of clusters as an input parameter. It only necessitates defining the minimum number of points (MinPts) or the neighborhood radius (ϵ-neighborhood of a point) required to form a cluster, thereby achieving clustering of arbitrarily shaped point clouds.

Figure 5.

Diagram divided into three panels labeled a, b, and c, illustrating a process for identifying and locating plant information points in cotton. Panel a shows locating layer threshold cutting of GP with colored dots and data layers. Panel b visualizes clustering plant points and associating them with plant information. Panel c demonstrates assigning and positioning information, including plant variety, repeated sample plot, and index number. Insets magnify data points at each stage, and color-coded data represent different categories. Text annotations explain each processing step.

Extraction of core points for plant localization. (a) Point cloud subsetting by height layers. (b) Clustering of point clouds within the localization area to obtain plant localization point clouds, with different plants distinguished by random colors. (c) Obtaining the mean point at a fixed height within the localization area as the plant localization point, enabling plant information association.

First, a localization height layer [Ll, Lh] is defined. From the cotton point cloud P, which has had ground and noise points removed, a subset of point clouds Call is extracted. This subset represents the root localization point cloud collection for all plants. Next, independent plant root point clusters are formed. Starting from an arbitrary point p in the point cloud set Call, its ϵ-neighborhood (denoted as (p)), as Equation 6 is defined as the set of all points q (including p) such that the distance between p and q is within the ϵ radius:

N(p)={qD|dist(p,q)ϵ} (6)

If the current neighborhood satisfies |N(q)|≥ MinPts, a cluster is initiated. Otherwise, point p is classified as a noise point. This process is repeated for every point in D. When the point cloud exhibits a relatively uniform distribution, it is not necessary to provide both MinPts and ϵ parameters simultaneously. Therefore, ϵ can be determined using the following formula:

ϵ=T·MinPts·Γ[(12)·n+1]mπn (7)

Where m represents the number of points, n is the dimensionality of the points, Γ is the Gamma function, X is the point cloud set, and T is the volume of the space formed by m points (Equation 8). The clustering result yields s point clusters: C1, C2,…,Cs. Each cluster corresponds to the root localization region of a single cotton plant. Finally, the mean point of cluster Ckat a fixed height Lhis calculated, serving as the localization point Apk (Equation 9) for cluster Ck:

T=i=1n{max (Xi)min (Xi)} (8)
Apk=(1|Ck|pCkxp,1|Ck|pCkyp,z¯0) (9)

Where, z¯0=Lh. This process uniquely defines each cotton plant in the population by a localization point Apk, which serves as its identification code. This identification code, together with the plant variety, plot number, plant index number, and spatial coordinates, constitutes the field information for each plant. Subsequently, a mapping relationship is established with the extracted phenotypic information.

2.4.3. Step 3: spatially constrained plant point cloud segmentation

Region growing is a classic algorithm in point cloud segmentation. Its main idea is to evaluate the similarity between seed points and neighboring points, classifying units with similar characteristics into one category. Its core steps involve seed point selection, similarity metric construction, and cluster scope definition. Envelope analysis of cotton plants reveals that the angle between their main stem and the outermost points of the leaf canopy consistently falls within a certain range. This study uses the point cloud set of the plant’s localization region as initial seed points. By calculating spatial constraint features based on plant height and plant architecture, it enables the growth of the population point cloud within the constrained space of each plant. This approach differs from previous region growing methods by innovatively linking the scope of region growing with the plant’s own characteristics such as height and architecture, thereby improving the algorithm’s efficiency and accuracy.

It primarily includes four steps: calculation of plant region height Rhk, definition of spatial constraint range Cyk, construction of multi-constraint growth rules, and correction of overlapping point cloud attribution. The pseudo-code of the point cloud instance segmentation step is presented in Algorithm 1.

Algorithm 1

Algorithm 1

1. Calculation of plant region height Rhk.

Centered at each plant localization point Apk, a cylindrical envelope with radius d is constructed. The point cloud Pk,small within this cylinder is then obtained, such that:

Rhk=max {zp|pPk,small} (10)

2. Definition of spatial constraint range Cyk.

Defining a spatial constraint range Cykfor each plant limits the growth scope of its point cloud, significantly reducing the number of point clouds within that growth range. This prevents overgrowth and minimizes erroneous segmentation of point clouds between adjacent plants.

Centered at the plant localization point Apk, and constrained by the angle between the plant’s main stem and the outermost envelope point, the spatial constraint range α for the plant point cloud is constructed as follows:

Cyk:{(xApk×x)2+(yApk×y)2rk2zzmin (11)

Where, rk= Rhk · tan(α) is the base radius of the spatial constraint, and zmin represents the minimum Z-coordinate of the point cloud within the region of interest. The pseudo-code for the calculation of plant positioning points and growth radius is shown in Algorithm 2.

Algorithm 2

Algorithm 2

3. Construction of multi-constraint growth rules.

We define three constraint metrics: the Euclidean distance constraint D(p,q) the normal vector similarity constraint N(p,q), and the spatial range constraint Y(q,Cyk). Their respective formulas are as follows:

D(p,q)={1||pq||<δ0otherwise (12)
N(p,q)=np·nq||np||·||nq||τN (13)
Y(q,Cyk)={1qCyk0otherwise (14)

Where, δ represents the distance threshold between growing points, ensuring the continuity of the point cloud space; τN is the cosine of the angle between normal vectors; and np and nq are the unit normal vectors of points p and q, respectively. For a point p in the queue, the growth condition for its neighboring point q is (D(p,q)=1)(N(p,q)τN)(Y(q,Cyk)=1).

4. Correction of overlapping point cloud attribution.

The overlapping regions of the constraint ranges invariably consist of point clouds from the marginal leaves of adjacent plants. The ownership of these intersection points is determined by the leaves within their respective neighborhood spaces. In a spatial context, this manifests as continuous local point cloud clusters, characterized by high density and consistent normal vectors within their neighborhoods. Extract the local neighborhood point set Nkq={pPk|||pq||<Rloc}for the intersection point q within the point cloud Pkof candidate plant k. Subsequently, filter out points belonging to different leaves based on normal vector consistency to obtain the fitting point set Fk,q. The RANSAC algorithm is then applied to Fk,qto fit a plane, whose equation is:

ak,qx+bk,qy+ck,qz+dk,q=0 (15)

Where (ak,q,bk,q,ck,q) is the unit normal vector of the plane, satisfying ak,q2+bk,q2+ck,q2=1. The intersection points are then verified for assignment based on local plane distance and curvature consistency. The distance dk,q from the intersection point q to this plane can be defined as:

dk,q=|ak,qxq+bk,qyq+ck,qzq+dk,q|ak,q2+bk,q2+ck,q2 (16)

Calculate the mean curvature μk,q and standard deviation σk,q of Fk,q. The curvature κq of q should satisfy |κqμk,q|γ1σk,q, while also satisfying the local planar distance dk,q<γ2. Where, γ1 and γ2 are empirical values. If these conditions are met, the intersection point q is assigned to plant k. The pseudo-code of the region growing step based on dynamic constraints is shown in Algorithm 3.

Algorithm 3

Algorithm 3

2.5. Phenotypic trait analysis

Cotyledonary Node Height (FCNH) in cotton is a crucial morphological indicator for assessing lodging resistance and early growth vigor in cotton seedlings. Excessively high FCNH increases the risk of plant lodging. Furthermore, studies have demonstrated a direct correlation between the number of main stem nodes (including cotyledonary nodes) and boll distribution and yield (Reddy et al., 2017). Its mathematical expression is:

FCNH=zfnzgd (17)

Where zfn represents the Z-coordinate of the cotyledonary node, and zgd represents the ground datum plane. The cotyledonary node is identified by first removing leaf sections from the cotton plant using plane fitting, then obtaining the intersection points of the main stem and branches through line fitting. The intersection point with the minimum Z-coordinate is designated as the cotyledonary node.

Plant height (PH) is defined as the vertical distance from the cotyledon node to the growing point at the apex of the main stem. It serves as a core morphological parameter reflecting the overall growth posture of the plant. PH is a crucial indicator for monitoring cotton growth and assessing lodging resistance (Reddy et al., 2017). Additionally, in breeding selection, consistency in plant height is a key criterion for evaluating hybrid cotton varieties. Its mathematical expression is:

PH=zmsvzgd (18)

Where, zmsv represents the Z-coordinate value of the plant’s highest point. Leaf area refers to the sum of the surface areas of individual leaves on a plant, significantly influencing crop photosynthetic efficiency and stress response. For pre-processed plant point clouds, leaf point clouds are first obtained through plane fitting. Subsequently, cluster segmentation is performed based on normal and curvature features. Poisson surface reconstruction is then applied to the clustered point clouds, and the surface area of the resulting Poisson surface represents the leaf area of that particular leaf.

2.6. Evaluation measure

The experiment evaluates the 3D modeling process in terms of image fidelity, perceptual quality and point cloud sampling. Peak Signal-to-Noise Ratio (PSNR) is based on the mean square error and is mainly implemented to quantify the signal fidelity of the reconstructed image with the mathematical formula:

PSNR(I,K)=10log10(MAXI2MSE) (19)
MSE=1m×ni=1mj=1n(I(i,j)K(i,j))2 (20)

Where, I(i,j) and K(i,j) respectively denote the pixel values of the original and reconstructed images. The Structural Similarity Index Measure (SSIM) takes brightness, contrast and structural similarity into account to assess whether the reconstructed scene has structural features similar to the real scene with the mathematical formula:

SSIM(x,y)=[l(x,y)]α·[c(x,y)]β·[s(x,y)]γ (21)

Where, x and y denote localized window blocks extracted from two distinct images. Specifically, l(x,y) quantifies the luminance similarity, c(x,y) characterizes the contrast similarity, and s(x,y) measures the structural similarity, with α, β, γ functioning as a weighting parameter. LPIPS (Learned Perceptual Image Patch Similarity) measures the difference between rendered images and real images in the depth feature space, better reflecting the perceptual quality of generated images than traditional metrics:

LPIPS(I,K)=l=1wl·Fl(I)Fl(K)22 (22)

Where wl represents the weighting layer, Fl denotes the l-th layer features extracted from the pre-trained network, and ||·||22 signifies the squared Euclidean distance between the corresponding features. To evaluate the uniformity of distribution in the reconstructed cotton population point cloud, the point cloud surface density (Equation 23) and volume density (Equation 24) metrics were selected:

Dsur=NAcs (23)
Dvol=NVcv (24)

where Acs is the surface area of the convex packet of the point cloud (mm2), Vcv is the volume of the convex packet (mm3), and N is the effective number of points.

In Cloudcompare software, plant point clouds were manually annotated from the group point cloud, with each plant assigned a distinct label. This serves as ground truth (GT) data for evaluating the accuracy of instance segmentation algorithms. In the evaluation of plant point cloud segmentation, if a cotton plant exists and is segmented from the point cloud, it is called true positive (TP); if a cotton plant does not exist but is segmented from the point cloud, or a plant exists, it is completely segmented and a part of the point cloud of the nearby cotton is also segmented into the plant, it is called false positive (FP); if a cotton is not segmented and assigned to a nearby cotton, or a cotton exists and is segmented, but the segmentation is incomplete, the part of the cotton point cloud is identified as the part of other plants, which is called false negative (FN). High TP values and low FN and FP values correspond to high accuracy. Recall rate (r), intersection-union ratio (IoU), precision rate (p) and F score are calculated, which are commonly used to evaluate the segmentation performance of point cloud instances (Tao et al., 2015).

2.7. Parameter sensitivity analysis and robustness verification

To ensure the generalization and robustness of the proposed instance segmentation method, this study follows a systematic parameter tuning and validation pipeline. All core algorithm parameters were optimized via grid search combined with manual visual assessment on a representative standardized experimental plot (VARf6REPa). This tuning set was used to identify parameter combinations that achieve an optimal balance between clustering completeness, boundary clarity, and other visual quality metrics and quantitative indicators.

The core parameters determined through tuning are listed in Table 2. These parameter values are set based on typical geometric scales of cotton plants in standardized breeding fields: the neighborhood parameter is smaller than the typical inter-plant spacing to distinguish adjacent plants, the distance threshold δ matches the branch diameter scale, and the normal vector threshold τNis used to capture continuous orientation of plant surfaces.

Table 2.

Core parameters for instance segmentation.

Parameter Range Optimized value
DBSCAN neighborhood ϵ 0.8 cm – 2.0 cm 1.2 cm
Minimum points MinPts 3 – 10 6
Initial cylinder radius d 2.0 cm – 4.0 cm 3.0 cm
Search cone angle α 20° – 50° 35°
Distance threshold δ 1.0 cm – 3.0 cm 1.5 cm
Normal vector threshold τN 0.70 – 0.95 0.85

The optimized parameter set was fixed and directly applied to eight additional independent trial plots covering different varieties and planting replicates for validation. The method achieved consistently high segmentation accuracy across all test plots (average F1-score of 91.32%). The results demonstrate that the parameter group tuned based on representative samples exhibits strong generalization capability in the standardized breeding field environment, and the proposed method is robust to common field variations such as plant density fluctuations and point cloud quality changes.

To evaluate the robustness of the method, we tested the impact of key parameters varying within ±20% of their nominal values on segmentation accuracy (mIoU). The results show that variations in all parameters within this range led to less than a 5% change in mIoU, indicating the method’s insensitivity to parameter settings. Furthermore, tests were conducted on subsets of data with different reconstruction qualities. Segmentation accuracy (both mIoU and F1-score) remained stable across all subsets, verifying the robustness of the method to common field variations.

3. Results

The experimental setup was equipped with an Intel Core i9-14900F processor, 128 GB of RAM, and an NVIDIA GeForce RTX 4090 24G graphics card. The system operated on Ubuntu 20.04, where COLMAP was employed for feature extraction and sparse reconstruction. The environment for 3D Gaussian Splatting and dense point cloud reconstruction was configured with CUDA 11.8 and PyTorch 2.1.1. Subsequent point cloud processing was also conducted on Ubuntu 20.04, primarily utilizing the PCL library, along with Python Open3D and NumPy for implementation.

3.1. 3D reconstruction of cotton plant populations

Three-dimensional modeling of cotton populations constitutes the foundational and paramount initial phase for the extraction of cotton population phenotypic traits. The experiment selected MVS, NERF, and PhenotypeAI methods to model cotton populations based on the constructed cotton field dataset. The experimental outcomes underwent qualitative scrutiny, focusing on the completeness of reconstruction for various cotton cultivar sample plots and the incidence of voids or gaps within plant organs. Figure 6 illustrates the three-dimensional point cloud models derived from the MVS, NeRF, and PhenotypeAI, taking the VARf6_REPa group data as an example. Table 3 provides a quantitative analysis, detailing the computational time expended for modeling and elucidating key performance metrics related to accuracy. At the population scale, all three methods—MVS, NeRF, and PhenotypeAI—achieved plant-level modeling of cotton. The red areas along the plant edges in the top view reveal significant artifact points in the MVS model. Delving into the scale of individual plant organs, panel (d) distinctly illustrates that the main stems reconstructed by both MVS and NeRF methods exhibit considerable voids, and in some instances, complete point cloud omissions. Conversely, the top-down perspective provided in panel (d) indicates that all three approaches attained commendable reconstruction of cotton leaf point clouds within the plant canopy. Consequently, the PhenotypeAI method exhibits less noise and more continuous plant organs in the point cloud reconstruction of cotton populations in the field.

Figure 6.

Three side-view plant point cloud reconstructions labeled (a) MVS, (b) NeRF, and (c) PhenotypeAI show decreasing main stem fractures from top to bottom, as annotated by contrasting red labels. Panel (d) compares zoomed-in details from each method, highlighting stem continuity and fracture differences with colored circles. Three small vertical views display the top-down perspective for each method, emphasizing completeness and segmentation clarity of plant structures.

Comparison of VARf6_REPa population modeling methods. (a) Group point cloud constructed using the MVS method; (b) Group point cloud constructed using the NeRF method; (c) Group point cloud constructed using the PhenotypeAI method; (d) Local enlargements and partial top-down views of group point clouds generated by the three methods.

Table 3.

Cotton population reconstruction accuracy analysis.

Varieties Method PSNR LPIPS SSIM Time/h Dsur Dvol
VARc94 MVS 9.3 485 267
NeRF 27.0 0.22 0.72 13.5 197 165
PhenotypeAI 30.0 0.20 0.84 1.1 394 260
VARf6 MVS 8.5 549 291
NeRF 27.8 0.24 0.80 14.2 220 183
PhenotypeAI 31.3 0.18 0.86 1.2 437 309
VARx36 MVS 9.1 622 348
NeRF 25.2 0.26 0.67 15.0 244 196
PhenotypeAI 32.1 0.17 0.87 1.4 475 334

Regarding image quality evaluation metrics, the PSNR value for the VARc94 variety reached 30.0 dB, representing an 11.1% improvement over the NeRF method. The SSIM metric achieved 0.84, a 16.7% increase over NeRF, indicating that PhenotypeAI holds a significant advantage in preserving structural similarity. For the LPIPS perceptual quality metric, PhenotypeAI reduced the value for the VARx36 variety to 0.17, a 34.6% decrease compared to NeRF, demonstrating that its reconstruction results better align with human visual perception characteristics. Multi-variety comparisons show that the VARx36 variety achieved a PSNR of 32.1 dB in the PhenotypeAI model, 6.7% and 4.2% higher than VARc94 and VARf6 respectively.

In terms of computational efficiency, the PhenotypeAI model’s training time metrics further highlight its engineering value. Point cloud modeling requires only 1.1-1.4 hours, representing a reduction of over 90% compared to NeRF. This efficiency gain stems from PhenotypeAI’s adaptive Gaussian sphere optimization, which concentrates computational resources on regions of high feature significance. This achieves a surface point density of 475 points per square centimeter across the VARx36 variety, marking a 94.7% improvement over NeRF.

The study evaluated the performance differences of three modeling methods—MVS, NeRF, and PhenotypeAI in the 3D reconstruction of cotton plant populations. Experimental data indicate that the PhenotypeAI method demonstrated significant advantages in terms of visual quality, computational efficiency, and model accuracy.

The plot point cloud of the sample area reconstructed using the PhenotypeAI method is shown in Figure 7. Based on this reconstructed point cloud, subsequent research on instance segmentation of cotton field plant point clouds will be conducted.

Figure 7.

Grid of nine false-color aerial images arranged in three rows and three columns, each showing plant plots with varying color patterns. Rows are labeled VARx36, VARf6, and VARc94; columns are labeled REPa, REPb, and REPc.

Cotton population reconstruction results. The varieties are VARx36, VARf6, and VARc94, with three experimental plots for each variety.

3.2. Point cloud processing and instance segmentation

The quantitative evaluation results of cotton population instance segmentation are shown in Tables 4, 5. Combined with the overall preprocessing effect of point cloud, the integrity of segmented plants and the characteristics of noise distribution shown in Figure 8, the qualitative analysis is carried out. The qualitative analysis takes the VARf6_REPa group as an example.

Table 4.

PhenotypeAI instance segmentation and optimization results.

Varieties REP Manually counted Divided FP FN TP r p F
VARc94 a 43 43 3 5 40 88.89 93.02 90.91
b 45 45 4 3 41 93.18 91.11 92.13
c 45 45 5 5 40 88.89 88.89 88.89
VARf6 a 44 44 3 4 41 91.11 93.18 92.13
b 43 43 3 4 40 90.91 93.02 91.95
c 45 45 4 3 41 93.18 91.11 92.13
VARx36 a 48 48 5 3 43 93.48 89.58 91.49
b 46 46 4 3 42 93.33 91.30 92.31
c 44 44 4 5 40 88.89 90.91 89.89

Table 5.

Time efficiency assessment of cotton population point cloud instance segmentation via PhenotypeAI.

Varieties VARc94 VARf6 VARx36
a b c a b c a b c
Ttol/s 56.22 57.18 60.94 58.12 56.40 60.15 62.09 61.38 58.42
Tavg/s 1.31 1.27 1.35 1.32 1.31 1.34 1.29 1.33 1.33

Figure 8.

Four-panel scientific figure showing a 3D point cloud of plants with data processing stages. Panel (a) displays the original multi-colored point cloud and an inset zooming on outlier points. Panel (b) shows preprocessing, highlights edges, and includes an inset. Panel (c) shows segmented single plants with ground points removed. Panel (d) assigns random colors to individual plants, each surrounded by a red bounding box.

VARf6_REPa population point cloud instance segmentation results: (a) is the reconstructed original population point cloud, (b) is the population point cloud after preprocessing and noise removal, (c) is the cotton population point cloud after ground removal, and (d) is the segmented plant point cloud, with different individual plants represented by random colors.

Table 4 records the evaluation results of the PhenotypeAI method on the segmentation accuracy of VARc94, VARf6 and VARx36 at the plant scale of each sample area of the three varieties. Among them, Manually calculated is the real number of cotton plants, Divided is the number of cotton plants detected by the algorithm, and the number of cotton plants in each sample area is correctly detected, mainly due to the fact that we determine the number and location of cotton plants through high-precision cotton positioning area.

Table 5 records the evaluation results of the instance segmentation efficiency of the PhenotypeAI method on the plant scale of VARc94, VARf6 and VARx36, which is the total instance segmentation time of the population point cloud in each sample area and the average segmentation time of each plant.

As shown in Table 6, this study conducted a comparative evaluation of the proposed point cloud segmentation method against classical point cloud segmentation methods on the collected cotton plant point cloud dataset. The results demonstrate that our proposed point cloud segmentation method achieves superior performance across all segmentation accuracy metrics, significantly outperforming classical point cloud segmentation methods. Our method successfully identified all plant instances, owing to the proposed point cloud localization regions. This approach reduces over-reliance on crop morphological features and point cloud density during segmentation, achieving a maximum F-score of 92.08%.

Table 6.

Comparison of results from different point cloud segmentation methods.

Method Varieties Manually counted Divided FP FN TP r p F
DBSCAN VARc94 133 79 41 36 38 51.35 48.10 49.67
VARf6 132 81 37 29 44 60.27 54.32 57.14
VARx36 138 75 43 35 32 47.76 42.67 45.07
Euclidean VARc94 133 87 39 42 48 53.33 55.17 54.24
VARf6 132 95 43 39 52 57.14 54.74 55.91
VARx36 138 93 40 43 53 55.21 56.99 56.08
HDBSCAN VARc94 133 112 21 20 91 81.98 81.25 81.61
VARf6 132 115 18 17 98 85.22 84.48 84.85
VARx36 138 118 22 19 99 83.90 81.82 82.85
Ours VARc94 133 133 12 13 121 90.30 90.98 90.64
VARf6 132 132 10 11 122 91.73 92.42 92.08
VARx36 138 138 13 11 125 91.91 90.58 91.24

We recorded the segmentation efficiency results of the comparative algorithms, as shown in Table 7. The Euclidean method required the longest processing time for group point cloud segmentation, while the DBSCAN method was the fastest. However, DBSCAN failed to fully capture all cotton plants within the region. HDBSCAN, a more advanced geometric clustering baseline method, can automatically determine the number of clusters and handle density variations, outperforming both fixed-threshold Euclidean clustering and DBSCAN in segmentation effectiveness. In areas with significant plant adhesion, although HDBSCAN could distinguish density differences, it still struggled to accurately segment interpenetrating branches. By dividing the total group segmentation time by the number of plants segmented by each algorithm, we calculated the average processing time per plant. Among the four algorithms, the localization and growth method proposed in this study, incorporating plant-level topological constraints and based on a continuous cylindrical model of the main stem, more effectively guided segmentation boundaries and demonstrated significant advantages in handling complex intersecting structures. Meanwhile, due to its superior ability to accurately detect individual plants within the group, it achieved the shortest average processing time per plant.

Table 7.

Comparison of computational time among different point cloud segmentation methods.

Method Ttol/s Tavg/s
VARc94 VARf6 VARx36 VARc94 VARf6 VARx36
DBSCAN 131.47 130.48 136.41 1.66 1.61 1.82
Euclidean 281.29 279.17 291.86 3.23 2.94 3.14
HDBSCAN 292.86 293.86 294.86 2.61 2.56 2.50
Ours 174.34 174.67 181.89 1.31 1.32 1.32

3.3. Phenotypic trait extraction

Figure 9 displays partial plant point clouds after instance segmentation. We obtained a total of 403 complete cotton plant point clouds, and a unique information system was established for each plant, primarily storing the cotton variety, replicate number, and Index Number within the entire sample. The analyzed phenotypic information will also be supplemented in the information system of the corresponding plant. This system uses plant localization points as a benchmark for spatial positioning and is effective throughout the entire growth cycle of the cotton.

Figure 9.

Nine 3D surface plots arranged in three rows labeled a, b, and c, each illustrating cotton plant structures for different varieties—VARX36, VAR6, and VARC94—across repeated sample plots and index numbers, with color gradients indicating measured data values.

Single plant point clouds after instance segmentation: (a) VARx36 variety, plant Index numbers 1-138; (b) VARf6 variety, plant Index numbers 139-270; (c) VARc94 variety, plant Index numbers 271-403.

Figure 10 presents the correlation analysis between the predicted and observed values of cotton cotyledon node height, plant height, and leaf area. The results demonstrate that extracting plant phenotypic traits based on 3D point cloud data achieves high accuracy: the coefficient of determination (R2) for cotyledon node height prediction is 0.90 (95% CI: 0.87–0.92), with a relative root mean square error (rRMSE) of 4.38% ± 0.51%. The R2 for plant height prediction is 0.91 (95% CI: 0.89–0.93), with an rRMSE of 5.69% ± 0.63%. For leaf area prediction, the R2 reaches 0.91 (95% CI: 0.88–0.94), with an rRMSE of 13.58% ± 1.22%. These statistics not only confirm the core accuracy of the method, but their accompanying confidence intervals and variability measures also quantify the stability of the estimates, providing a more robust basis for the reproducibility and comparability of the results.

Figure 10.

Three scatter plots labeled (a), (b), and (c) compare calculated versus measured values for three cultivars, VARc94, VARf6, and VARx36. Each plot displays a regression line, equation, R squared value, and root mean square error.

Correlation analysis between predicted and measured cotyledon node height, plant height, and leaf area in cotton. (a) Linear regression between predicted and measured cotyledon node height. (b) Linear regression between predicted and measured plant height. (c) Linear regression between predicted and measured leaf area.

Figure 11 the distribution ranges and error bars for the acquired traits of cotyledon node height, plant height, and leaf area. The distribution of cotyledon node height was relatively stable across varieties, whereas plant height and leaf area exhibited significant variation in numerical distribution, reflecting the varietal characteristics of cotton. These results indicate substantial differences in growth among cotton varieties. And the PhenotypeAI method proposed in this paper can effectively obtain these differences.

Figure 11.

Three box plot panels compare plant height (a), cotyledon node height (b), and leaf area (c) among three varieties (VARc94, VARf6, VARx36) across blocks a, b, and c, with color legends and labeled axes.

Distribution ranges and error bars of cotyledon node height, plant height, and leaf area traits obtained from cotton plants. (a) Distribution range and error bars for cotyledon node height; (b) Distribution range and error bars for plant height; (c) Distribution range and error bars for leaf area. In the figure, blue denotes the VARc94 cultivar, orange represents VARf6, and green indicates VARx36; varying shades are used to distinguish different sampling plots within each cultivar.

Figure 12 Correlation Analysis Among Cotyledon Node Height, Plant Height, and Leaf Area Traits Across Cotton Varieties. The analysis results in the figure below indicate that cotyledon node height and plant height traits exhibit strong correlations among cotton varieties, while the correlation between plant height and leaf area traits is the weakest. The correlation between cotyledon node height and leaf area traits is also the weakest. This aligns with the previously obtained cotyledon node distribution results, primarily because the differences in cotyledon node height traits among cotton varieties at this stage are relatively small, whereas the differences in plant height and leaf area traits are more pronounced. .

Figure 12.

Four correlation matrices for plant traits are shown in heatmap format, each for different conditions: Overall, VARc94, VARf6, and VARx36. Traits include plant height, cotyledon node height, and leaf area. Color intensity represents correlation strength, with darker red indicating higher positive correlation and light blue indicating lower or negative correlation, as shown by the color scale from zero to one. Each heatmap displays similar patterns of strong positive correlations between the three measured traits. Panel labels (a), (b), (c), and (d) correspond to Overall, VARc94, VARf6, and VARx36, respectively.

Correlation analysis among cotyledonary node height, plant height, and leaf area traits in different cotton cultivars. (a) Correlation analysis among cotyledonary node height, plant height, and leaf area for all cotton plants; (b) Correlation analysis among these traits for VARc94 cultivar; (c) Correlation analysis for VARf6 cultivar; (d) Correlation analysis for VARx36 cultivar.

Table 8 Time consumption analysis for cotton trait extraction using PhenotypeAI. The table records the total extraction time for cotyledonary node height, plant height, and leaf area traits across all cultivars and sampling plots, as well as the average extraction time per individual plant for these traits. Overall, the extraction time for cotyledonary node height, plant height, and leaf area in each sampling plot ranged from a minimum of 34.72 seconds to a maximum of 39.79 seconds, while the extraction time per individual plant ranged from 0.85 to 0.93 seconds.

Table 8.

Time efficiency assessment of cotton phenotypic trait extraction via PhenotypeAI.

Varieties VARc94 VARf6 VARx36
a b c a b c a b c
T eta /s 35.34 38.69 37.65 34.86 34.72 35.93 39.79 37.46 35.61
T avg /s 0.82 0.86 0.84 0.79 0.81 0.80 0.83 0.81 0.81

4. Discussion

4.1. PhenotypeAI markedly outperforms sfm-mvs and nerf methods in reconstruction detail

Compared with SFM-MVS and NeRF methods, PhenotypeAI achieved a peak signal-to-noise ratio (PSNR) of 30.0 dB in reconstruction, producing point clouds with the lowest noise level while exhibiting the highest completeness and fineness of plant point clouds. This is crucial for enhancing the success rate of point cloud instance segmentation and the accuracy of phenotypic information extraction. The core advantage of PhenotypeAI lies in its adoption of an explicit discrete representation based on 3D Gaussian Splatting (3DGS). This representation describes a scene as a collection of Gaussian ellipsoids with optimizable parameters, where the scale and orientation of local geometry are directly controlled through covariance matrices. For thin structures such as cotton stems, 3DGS can accurately capture their morphology by arranging slender and directionally strong Gaussian primitives, avoiding the detail blurring caused by the smoothness prior and numerical integration in implicit continuous representations like NeRF. As a result, PhenotypeAI enables more faithful fine-scale geometric reconstruction. In complex cotton canopy scenarios, PhenotypeAI demonstrates significant advantages, requiring only 1.1–1.4 hours to complete point cloud modeling, which is over 90% faster than NeRF-based algorithms, thereby highlighting its potential to advance plant phenotyping analysis.

4.2. PLZ for the management of field information and phenotypic traits of individual cotton plants

The accurate localization of plants within a population is crucial for managing phenotypic traits, and a plant localization zone enables the long-term, time-series tracking of individual plant traits. This study leverages the independent growth characteristics of cotton plant stems. By extracting the point cloud of the plant community after ground removal at a specific height layer, root points of individual plants are identified through density-based clustering, and the positioning points for each plant are calculated. This localization point can establish a long-term mapping relationship between the plant’s positional information and its phenotypic data, thereby aiding in the management of plant-level phenotypic information. The proposed method is particularly suitable for high-standard farmlands implementing precision agronomic practices, where interference from robust weeds and other complex factors is minimized. In future work, we plan to incorporate stem diameter, branching angle, and spectral features at specific heights to enhance the generalization capability of the method in complex, non-standard scenarios.

4.3. PhenotypeAI has the advantages of high efficiency and low computational cost in population point cloud instance segmentation

Instance segmentation of the population point cloud is an important component for extracting plant traits. In this study, the segmentation accuracy for cotton plant instances reached 88.89% 93.18%, with completion time reduced to just 56.22 seconds, significantly improving efficiency. Additionally, there are several deep learning-based point cloud segmentation methods (Chen et al., 2025). manually labeled cotton point clouds using CloudCompare software and employed the 3D deep neural network PointNet++ for population point cloud segmentation to extract individual cotton plants and the population, achieving a cotton detection rate of R = 86.3%. Although deep learning is currently a very popular method for point cloud segmentation, it relies on large amounts of data for training. Manually labeling this data is error-prone, labor-intensive, and costly. Current deep learning methods also require downsampling the point cloud to a very low resolution, which leads to the loss of many potential geometric features of small organs and plants. The PhenotypeAI method proposed in this study does not require data labeling and training, which saves time and cost, and it exhibits good segmentation performance. Its high efficiency and low computational cost provide technical support for the precise analysis of field phenotypic traits for efficient cotton breeding.

4.4. Future work

This study demonstrates that the integration of smartphone-based imaging with Gaussian splatting rendering techniques enables efficient acquisition of point cloud data in cotton fields, effectively accomplishes instance segmentation of cotton populations, and accurately estimates cotyledon node height, plant height, and leaf area of cotton plants. Specifically, the PhenotypeAI methodology achieved a PSNR of 30.0 dB in cotton field reconstruction, with instance segmentation accuracy reaching a minimum of 88.89% and an extraction rate of 100%. In cotton phenotypic trait extraction, plant height exhibited the highest accuracy and lowest error. The cotyledon node height and leaf area traits also satisfied the precision requirements essential for breeding applications. Additionally, we propose an effective smartphone-based method for collecting point cloud data in cotton breeding fields, generating currently scarce point cloud datasets for cotton fields and individual plants. Final experimental results demonstrate that our PhenotypeAI method enables high-throughput, low-cost outdoor cotton reconstruction, instance segmentation, and phenotypic information extraction. This provides technical support for cotton growth monitoring, yield prediction, precision pesticide application, and scientific management. The core idea of this method is unsupervised instance segmentation based on geometric localization of the main stem. For other crops with distinct main stems or branches, such as corn and tomatoes, this workflow is directly transferable. For crops without obvious upright main stems, such as Chinese cabbage and kale, the adaptability accuracy of the current method will significantly decrease. In future research, we will prioritize robust reconstruction in dynamic scenarios and strive to advance the method from standard breeding trial fields to applications in unstructured cotton fields.

5. Conclusion

This study successfully developed an innovative method named PhenotypeAI. This technology enables high-precision, low-cost 3D reconstruction of cotton plant populations under field conditions using video footage captured by smartphones. It simultaneously supports point cloud instance segmentation, thereby extracting key phenotypic traits from individual plants. This approach fully demonstrates the immense potential of combining portable smartphone devices with 3D point cloud technology for acquiring cotton phenotypic data in the field. We conducted field trials at a modern cotton breeding base to validate the proposed method. Experimental data demonstrated that point cloud reconstruction was achieved using smartphone-captured canopy images of cotton fields, with a peak signal-to-noise ratio (PSNR) of 30.0 dB and a point cloud density of 394 points per square centimeter. The average accuracy of instance segmentation at the population level reached 91.35%, successfully constructing 3D point cloud models for 403 high-quality cotton plants. The average processing time per plant was 1.3 seconds. Analysis of key plant traits-FCNH (first cotyledon node height), PH (plant height), and LA (leaf area)-exhibited correlation coefficients (R2) of 0.90, 0.91, and 0.91, respectively, with relative root mean square errors (rRMSE) of 4.38%, 5.69%, and 13.58%. Individual plant analysis completed in as little as 0.79 seconds. Overall, this study provides a low-cost solution for high-throughput field phenotyping of cotton, offering significant value for smart breeding, growth monitoring, and scientific field management. This method holds promise as a technical benchmark for modern breeding and efficient cultivation practices.

Funding Statement

The author(s) declared that financial support was received for this work and/or its publication. This research was supported by the Natural Science Foundation of Xinjiang Uygur Autonomous Region Outstanding Youth Science Fund project ‘Research on high-throughput phenotypic analysis and control of mobile robots considering cotton growth dynamics’ (2024D01E02) and the National Natural Science Foundation of China project ‘Research on bionic vision and adaptive grasping method of Xinjiang long-staple cotton picking robot’ (52265003).

Footnotes

Edited by: Talha Meraj, Atlantic Technological University, Ireland

Reviewed by: Ajay Kumar Patel, Saint Louis University, United States

Negin Harandi, Ghent University, Belgium

Data availability statement

The original contributions presented in the study are included in the article/supplementary material. Further inquiries can be directed to the corresponding authors.

Author contributions

BL: Investigation, Data curation, Formal analysis, Writing – review & editing, Methodology, Writing – original draft. XL: Data curation, Methodology, Conceptualization, Funding acquisition, Writing – review & editing, Supervision. JY: Investigation, Data curation, Writing – original draft. ZL: Writing – review & editing, Data curation. XTZ: Data curation, Supervision, Writing – original draft. XJZ: Methodology, Conceptualization, Project administration, Supervision, Writing – review & editing.

Conflict of interest

The author(s) declared that this work was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declared that generative AI was not used in the creation of this manuscript.

Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

  1. Abdurakhimov S., Sultanov U., Tukhtamishev M. (2023). Development of the mineral fertilizer application system for cotton yards under different soil-climatic conditions of Uzbekistan. E3S Web of Conferences 389, 03056. doi:  10.1051/e3sconf/202338903056, PMID: 41430790 [DOI] [Google Scholar]
  2. Bao Y., Ding T., Huo J., Liu Y., Li Y., Li W., et al. (2025). 3D gaussian splatting: Survey, technologies, challenges, and opportunities. IEEE Transactions on Circuits and Systems for Video Technol. 35, 6832–6852. doi:  10.1186/s13595-022-01143-x, PMID: 41933398 [DOI] [Google Scholar]
  3. Bian L., Zhang H., Ge Y., Čepl J., Stejskal J., El-Kassaby Y. A. (2022). Closing the gap between phenotyping and genotyping: Review of advanced, image-based phenotyping technologies in forestry. Ann. Forest. Sci. 79, 22. doi:  10.1186/s13595, PMID: 41933398 [DOI] [Google Scholar]
  4. Chawade A., Van Ham J., Blomquist H., Bagge O., Alexandersson E., Ortiz R. (2019). High-throughput field-phenotyping tools for plant breeding and precision agriculture. Agronomy 9, 258. doi:  10.3390/agronomy9050258, PMID: 41725453 [DOI] [Google Scholar]
  5. Chen C., Jia Y., Zhang J., Yang L., Wang Y., Kang F. (2024). Development of a 3D point cloud reconstruction-based apple canopy liquid sedimentation model 451, 142038. doi:  10.1016/j.jclepro.2024.142038, PMID: 41936479 [DOI] [Google Scholar]
  6. Chen X., Wen S., Zhang L., Lan Y., Ge Y., Hu Y., et al. (2025). A calculation method for cotton phenotypic traits based on unmanned aerial vehicle LiDAR combined with a three-dimensional deep neural network. Comp. Electronics in Agriculture 230, 109857. doi:  10.1016/j.compag.2024.109857, PMID: 41936479 [DOI] [Google Scholar]
  7. Cui D., Liu P., Liu Y., Zhao Z., Feng J. (2025). Automated Phenotypic analysis of mature soybean using multi-view stereo 3D reconstruction and point cloud segmentation. Agriculture 15, 175. doi:  10.3390/agriculture15020175, PMID: 41725453 [DOI] [Google Scholar]
  8. Digumarti S. T., Nieto J., Cadena C., Siegwart R., Beardsley P. (2018). Automatic segmentation of tree structure from point cloud data. IEEE Robotics and Automation Letters 3, 3043–3050. doi:  10.1109/LRA.2018.2849499, PMID: 41116384 [DOI] [Google Scholar]
  9. Dipta B., Sood S., Devi R., Bhardwaj V., Mangal V., Thakur A. K., et al. (2023). Digitalization of potato breeding program: Improving data collection and management. Heliyon 9, e12974. doi:  10.1016/j.heliyon.2023.e12974, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
  10. Dong Y., Zhang Y., Du X., Wang H., Li Q., Shen Y., et al. (2025). Field-scale single-plant sunflower head detection and geometric parameters measurement integrating multi-modal UAV data, deep learning and point cloud analysis. Industrial Crops and Prod. 237, 122225. doi:  10.1016/j.indcrop.2025.122225, PMID: 41936479 [DOI] [Google Scholar]
  11. Elnashef B., Filin S., Lati R. N. (2019). Tensor-based classification and segmentation of three-dimensional point clouds for organ-level plant phenotyping and growth analysis. Computers and Electronics in Agriculture 156, 51–61. doi:  10.1016/j.compag.2018.10.036, PMID: 41936479 [DOI] [Google Scholar]
  12. Fu X., Jiang D. (2022). “ Chapter 16 - High-throughput phenotyping: The latest research tool for sustainable crop production under global climate change scenarios,” in Sustainable Crop Productivity and Quality Under Climate Change. Eds. Liu F., Li X., Hogy P., Jiang D., Brestic M., Liu B. ( Academic Press; ), 313–381. doi:  10.1016/B978-0-323-85449-8.00003-8, PMID: [DOI] [Google Scholar]
  13. Gao R., Cui S., Xu H., Kong Q., Su Z., Li J. (2023). A method for obtaining maize phenotypic parameters based on improved QuickShift algorithm. Computers and Electronics in Agriculture 214, 108341. doi:  10.1016/j.compag.2023.108341, PMID: 41936479 [DOI] [Google Scholar]
  14. Gu Y., Wang Y., Wu Y., Warner T. A., Guo T., Ai H., et al. (2024). Novel 3D photosynthetic traits derived from the fusion of UAV LiDAR point cloud and multispectral imagery in wheat. Rem. Sen. Environ. 311, 114244. doi:  10.1016/j.rse.2024.114244, PMID: 41936479 [DOI] [Google Scholar]
  15. He S., Li X., Chen M., Xu X., Tang F., Gong T., et al. (2024). Crop HTP technologies: Applications and prospects. Agriculture 14, 723. doi:  10.3390/agriculture14050723, PMID: 41725453 [DOI] [Google Scholar]
  16. He W., Ye Z., Li M., Yan Y., Lu W., Xing G. (2023). Extraction of soybean plant trait parameters based on SfM-MVS algorithm combined with GRNN. Front. Plant Sci. 14, 1181322. doi:  10.3389/fpls.2023.1181322, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
  17. Iqbal J., Xu R., Halloran H., Li C. (2020). Development of a multi-purpose autonomous differential drive mobile robot for plant phenotyping and soil sensing. Electronics 9, 1550. doi:  10.3390/electronics9091550, PMID: 41725453 [DOI] [Google Scholar]
  18. J R., Nidamanuri R. R. (2024). Deep learning-based prediction of plant height and crown area of vegetable crops using LiDAR point cloud. Scientific Rep. 14, 14903. doi:  10.1038/s41598-024-65322-8, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
  19. Jiang L., Rodriguez-Sanchez J., Snider J. L., Chee P. W., Fu L., Li C. (2025). Mapping of cotton bolls and branches with high-granularity through point cloud segmentation. Plant Methods 21, 66. doi:  10.1186/s13007-025-01375-8, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
  20. Kerbl B., Kopanas G., Leimkuehler T., Drettakis G. (2023). 3D gaussian splatting for real-time radiance field rendering. ACM Trans. Graph (TOG) 42, 1–14. doi:  10.1145/3592433, PMID: 41889584 [DOI] [Google Scholar]
  21. Li X., Han Y., Wang G., Feng L., Wang Z., Yang B., et al. (2020). Response of cotton fruit growth, intraspecific competition and yield to plant density. Euro. J. Agro. 114, 125991. doi:  10.1016/j.eja.2019.125991, PMID: 41936479 [DOI] [Google Scholar]
  22. Li X., Liu B., Shi Y., Xiong M., Ren D., Wu L., et al. (2024). Efficient three-dimensional reconstruction and skeleton extraction for intelligent pruning of fruit trees. Computers and Electronics in Agriculture 227, 109554. doi:  10.1016/j.compag.2024.109554, PMID: 41936479 [DOI] [Google Scholar]
  23. Li D., Wei Y., Zhu R. (2023). A comparative study on point cloud down-sampling strategies for deep learning-based crop organ segmentation. Plant Methods 19, 124. doi:  10.1186/s13007-023-01099-7, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
  24. Liang Z., Li X., Wang G., Wu F., Zou X. (2025). Palm vision and servo control strategy of tomato picking robot based on global positioning. Computers and Electronics in Agriculture 237, 110668. doi:  10.1016/j.compag.2025.110668, PMID: 41936479 [DOI] [Google Scholar]
  25. Ma X., Wei B., Guan H., Cheng Y., Zhuo Z. (2024). A method for calculating and simulating phenotype of soybean based on 3D reconstruction. Euro. J. Agro. 154, 127070. doi:  10.1016/j.eja.2023.127070, PMID: 41936479 [DOI] [Google Scholar]
  26. Mildenhall B., Srinivasan P. P., Tancik M., Barron J. T., Ramamoorthi R., Ng R. (2021). Nerf: Representing scenes as neural radiance fields for view synthesis. In Computer vision – ECCV 2020 (pp. 405–421). Springer International Publishing; 65, 99–106. doi:  10.48550/arXiv.2003.08934, PMID: 41363103 [DOI] [Google Scholar]
  27. Mirande K., Godin C., Tisserand M., Charlaix J., Besnard F., Hétroy-Wheeler F. (2022). A graph-based approach for simultaneous semantic and instance segmentation of plant 3D point clouds. Front. Plant Sci. 13, 1012669. doi:  10.3389/fpls.2022.1012669, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
  28. Peng H., Guo S., Zou X., Wang H., Xiong J., Liang Q. (2025). UAVO-NeRF: 3D reconstruction of orchards and semantic segmentation of fruit trees based on neural radiance field in UAV images. Computers and Electronics in Agriculture 237, 110631. doi:  10.1016/j.compag.2025.110631, PMID: 41936479 [DOI] [Google Scholar]
  29. Pepe M., Fregonese L., Crocetto N. (2022). Use of SfM-MVS approach to nadir and oblique images generated throught aerial cameras to build 2.5 D map and 3D models in urban areas. Geocarto Intern. 37, 120–141. doi:  10.1080/10106049.2019.1700558, PMID: 41909888 [DOI] [Google Scholar]
  30. Reddy K. R., Brand D., Wijewardana C., Gao W. (2017). Temperature effects on cotton seedling emergence, growth, and development. Agro. J. 109, 1379–1387. doi:  10.2134/agronj2016.07.0439 [DOI] [Google Scholar]
  31. Saeed F., Sun J., Ozias-Akins P., Chu Y. J., Li C. C. (2023). “ PeanutNeRF: 3D radiance field for peanuts,” in 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW). 6254–6263. doi:  10.1109/CVPRW59228.2023.00665, PMID: [DOI] [Google Scholar]
  32. Tancik M., Weber E., Ng E., Li R., Yi B., Wang T., et al. (2023). “ Nerfstudio: A modular framework for neural radiance field development,” in ACM SIGGRAPH 2023 Conference Proceedings. Siggraph ’23. doi:  10.1145/3588432.3591516, PMID: [DOI] [Google Scholar]
  33. Tang Y., Chen M., Wang C., Luo L., Li J., Lian G., et al. (2020). Recognition and localization methods for vision-based fruit picking robots: A review. Front. Plant Sci. 11, 510. doi:  10.3389/fpls.2020.00510, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
  34. Tao S., Wu F., Guo Q., Wang Y., Li W., Xue B., et al. (2015). Segmenting tree crowns from terrestrial and mobile LiDAR data by exploring ecological theories. ISPRS J. Photogrammetry and Rem. Sens. 110, 66–76. doi:  10.1016/j.isprsjprs.2015.10.007, PMID: 41936479 [DOI] [Google Scholar]
  35. Wang Z., Xiao S., Miao Z., Liu R., Chen H., Wang Q., et al. (2025. a). P3DFusion: A cross-scene and high-fidelity 3D plant reconstruction framework empowered by vision foundation models and 3D Gaussian splatting. Plant Phenomics 171, 127811. doi:  10.1016/j.eja.2025.127811, PMID: 41936479 [DOI] [Google Scholar]
  36. Wang H., Ye Z., Zhang Q., Wang M., Zhou G., Wang X., et al. (2025. b). RsegNet: an advanced methodology for individual rubber tree segmentation and structural parameter extraction from UAV liDAR point clouds. European J. Agro. 7, 100090. doi:  10.1016/j.plaphe.2025.100090, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
  37. Yu G., Ma B., Zhang R., Xu Y., Lian Y., Dong F. (2025). CPD-YOLO: A cross-platform detection method for cotton pests and diseases using UAV and smartphone imaging. Industrial Crops and Products 234, 121515. doi:  10.1016/j.indcrop.2025.121515, PMID: 41936479 [DOI] [Google Scholar]
  38. Zhang D., Gajardo J., Medic T., Katircioglu I., Boss M., Kirchgessner N., et al. (2025). “ Wheat3DGS: in-field 3D reconstruction, instance segmentation and phenotyping of wheat heads with gaussian splatting,” in Proceedings of the Computer Vision and Pattern Recognition Conference (CVPR) Workshops. 5369–5379. doi:  10.1109/CVPRW67362.2025.00533, PMID: [DOI] [Google Scholar]
  39. Zou R., Zhang Y., Chen J., Li J., Dai W., Mu S. (2023). Density estimation method of mature wheat based on point cloud segmentation and clustering. Comp. Elec. Agricult. 205, 107626. doi:  10.1016/j.compag.2023.107626, PMID: 41936479 [DOI] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

The original contributions presented in the study are included in the article/supplementary material. Further inquiries can be directed to the corresponding authors.


Articles from Frontiers in Plant Science are provided here courtesy of Frontiers Media SA

RESOURCES