Next Article in Journal
Improvement of the YOLOv8 Model in the Optimization of the Weed Recognition Algorithm in Cotton Field
Previous Article in Journal
Socioeconomic Disparities in Urban Forest Diversity and Structure in Green Areas of Santiago de Chile
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Research on Segmentation Method of Maize Seedling Plant Instances Based on UAV Multispectral Remote Sensing Images

1
School of Surveying and Land Information Engineering, Henan Polytechnic University, Jiaozuo 454000, China
2
Key Laboratory of Mine Spatio-Temporal Information and Ecological Restoration, Ministry of Natural Resources, Henan Polytechnic University, Jiaozuo 454000, China
*
Author to whom correspondence should be addressed.
Plants 2024, 13(13), 1842; https://doi.org/10.3390/plants13131842
Submission received: 11 April 2024 / Revised: 14 June 2024 / Accepted: 2 July 2024 / Published: 4 July 2024
(This article belongs to the Section Plant Modeling)

Abstract

The accurate instance segmentation of individual crop plants is crucial for achieving a high-throughput phenotypic analysis of seedlings and smart field management in agriculture. Current crop monitoring techniques employing remote sensing predominantly focus on population analysis, thereby lacking precise estimations for individual plants. This study concentrates on maize, a critical staple crop, and leverages multispectral remote sensing data sourced from unmanned aerial vehicles (UAVs). A large-scale SAM image segmentation model is employed to efficiently annotate maize plant instances, thereby constructing a dataset for maize seedling instance segmentation. The study evaluates the experimental accuracy of six instance segmentation algorithms: Mask R-CNN, Cascade Mask R-CNN, PointRend, YOLOv5, Mask Scoring R-CNN, and YOLOv8, employing various combinations of multispectral bands for a comparative analysis. The experimental findings indicate that the YOLOv8 model exhibits exceptional segmentation accuracy, notably in the NRG band, with bbox_mAP50 and segm_mAP50 accuracies reaching 95.2% and 94%, respectively, surpassing other models. Furthermore, YOLOv8 demonstrates robust performance in generalization experiments, indicating its adaptability across diverse environments and conditions. Additionally, this study simulates and analyzes the impact of different resolutions on the model’s segmentation accuracy. The findings reveal that the YOLOv8 model sustains high segmentation accuracy even at reduced resolutions (1.333 cm/px), meeting the phenotypic analysis and field management criteria.

1. Introduction

Maize, a traditional staple crop [1], is characterized by its diverse attributes, serving as an essential source of both feed and industrial raw materials. It is widely distributed worldwide [2]. Monitoring its growth process is crucial in agricultural production to increase yield and optimize cultivation practices.
However, traditional agricultural monitoring methods are constrained by labor and time costs, making it difficult to efficiently monitor large agricultural fields [3,4]. With the rapid advancements in imaging technology and artificial intelligence, deep learning techniques employing visible light images have been applied to the detection of crop seedlings [5]. The University of Alaska, USA, developed an advanced benchtop phenotyping platform [6] enabling the automated high-throughput testing of stationary plants with precise control over temperature, humidity, and light intensity. The DroughtSpotter [7], produced by PhenoSpex in the Netherlands, and the WPScan [8], produced by WE PROVE SOLUTIONS (WPS), a phenotypic information collection platform, can extract parameters of plant transpiration dynamics and growth status at various stages through image collection. Compared with the above platforms, unmanned aerial vehicles (UAVs), which are not affected by ground conditions and can efficiently collect high-resolution images, have become a popular choice for crop detection research in recent years [9,10,11,12,13]. In agricultural monitoring, UAV-mounted multispectral sensors yield more accurate information in the infrared band compared to visible imaging, offering farmers and agronomists valuable insights into plant conditions. Relative to hyperspectral sensors, multispectral sensors are more cost-effective and lighter, facilitating their integration into UAVs with smaller payloads and broadening the application range of various UAV types in agriculture [14,15,16]. Zhao [17] et al. used UAV multispectral imagery to make a diagnosis of nitrogen nutrition index (NNI) in cotton. Hu [18] explored the potential of UAV technology in rapidly producing maps of leaf chlorophyll content (LCC) and fractional vegetation cover (FVC) from maize canopy orthophotos to evaluate maize maturity. Han [19] identified a strong correlation between the maize height estimated from UAV imagery and the actual ground truth height. While existing studies have analyzed maize biomass, chlorophyll content, and height, research on the accurate segmentation and leaf area analysis of individual maize seedlings remains scant.
With the rapid advancement of computer vision and deep learning technologies, image-based plant instance segmentation has emerged as a critical research tool, offering enhanced accuracy for agricultural production [20]. Turgut [21] used deep learning architecture based on attention mechanism to achieve plant organ segmentation by extracting contextual features and performing feature propagation to process point regions in a hierarchical manner. Lu [22] used UAV aerial images to obtain information such as canopy area, canopy width, location, etc., and proposed an unsupervised image segmentation method for the fast acquisition of fruit tree canopies under natural lighting conditions. Ren [23] first proposed a recurrent neural network-based recurrent instance segmentation algorithm to achieve leaf segmentation, but the final segmentation effect was very unsatisfactory; then, it was improved by using Conditional Random Fields (CRFs) [24] as the post-processing step of the network: the leaf segmentation effect was improved, and the evaluation metrics symmetric best dice (SBD) [25] was improved by about 10%. Yin [26] proposed the use of a watershed algorithm to complete the segmentation of individual leaves in an image, which in turn enables leaf counting. Zhang et al. [27] employed an unmanned aerial system to gather cabbage germplasm resources, achieving individual cabbage segmentation and enabling the estimation of its width, length, and relative chlorophyll content. Based on a comprehensive review of the existing literature, there is currently a lack of specialized research on the instance segmentation of early-stage maize seedlings. In addition, existing algorithms are not fully applicable to early-stage maize detection tasks.
The application of deep learning techniques necessitates a substantial volume of trainable data to fully leverage the sophisticated feature extraction and image analysis capabilities of deep convolutional neural networks [28,29]. Training samples for image instance segmentation necessitate pixel-level mask information, consequently escalating the cost of manual annotation. This limitation hampers the accurate representation of the plant’s phenotypic information. Qiang [30] achieved the segmentation of green leafy vegetable instances using deep convolutional neural networks; however, the model’s training demands extensive pixel-level mask-labeled data, with labeling often requiring several hours per image, thereby impeding the practical application of these algorithms. Bearman [31] et al. analyzed sample labeling methods, finding pixel-level annotation of an image, on average containing 2.8 objects, requires approximately 4 min, with the time cost for annotating single plant trait features being two to three times higher. To mitigate the manual annotation costs, Zhao introduced a weak label generation method using bounding boxes [32], facilitating the high-precision segmentation of maize seedling images at a reduced annotation time of approximately 2.5 min/sheet. While these methods offer effective solutions for image segmentation, their efficacy in processing field-specific images of young maize seedlings falls short of efficiency requirements. Furthermore, the volume of collected data samples is insufficient to comprehensively represent the diverse scenarios and conditions encountered in field environments.
The YOLOv8 algorithm introduces new features and enhancements that build on the success of previous You Only Look Once (YOLO) versions, with faster detection speed and higher accuracy, providing more advanced technical support for image processing. In contemporary research, the applications of YOLOv8 extend across diverse domains, such as automated waste sorting and detection [33], gravel soil mixture homogeneity assessment [34], pest field management [35], and the real-time detection of high-speed railway components [36], among others. However, despite YOLOv8’s impressive achievements in numerous fields, there is a notable absence of detailed research on applying the YOLOv8 model to young maize.
To date, significant advancements have been made in crop instance segmentation using UAV multispectral technology; however, the research in this field continues to confront numerous challenges. (1) The first is the lack of representative example segmentation datasets. Training samples for image instance segmentation necessitate pixel-level mask information, substantially increasing the cost of manual annotation, particularly in plant phenotyping. The diversity and complexity of images, alongside the variability of plant morphology and large sample sizes, make acquiring high-quality, representative datasets exceedingly challenging. (2) Another challenge is the limited applicability of instance segmentation algorithms. Not all existing algorithms are suitable for early-stage maize detection. Owing to the small size of early-stage maize plants and differences in leaf morphology and texture relative to mature plants, some algorithms demonstrate reduced accuracy and robustness in these tasks. Furthermore, most of the existing instance segmentation algorithms are optimized for outdoor scenes and mature crops, and they are ineffective in recognizing and segmenting the small scale, variable growth state and complex environmental disturbances of juvenile maize in farmland, which makes it difficult to meet the demand for the intelligent management of large-area maize farmland and achieve fine management.
In response to the aforementioned challenges, this paper presents the following principal contributions. (1) The first is the construction of an early-stage maize multispectral instance segmentation dataset. Utilizing the Segment Anything Model (SAM) [37] enabled the semi-automatic and efficient annotation of early-stage maize seedlings, thus creating a representative and high-quality dataset for maize seedling instance segmentation. (2) The second is evaluation of the performance of multiple mainstream instance segmentation algorithms on the constructed dataset. In addition, a special performance analysis was conducted for the overdense areas in corn planting to help farmers make more scientific field management decisions, optimize crop growth conditions and ultimately improve corn yield and quality.

2. Materials and Methods

The eastern region of Nanzhang Village, Shanyang District, Jiaozuo City, Henan Province, was selected as the research area (35°9′ N, 113°14′ E). The area is characterized by flat terrain, fertile soil, abundant water resources, and favorable agricultural production conditions. As a key contributor to China’s grain production, Henan Province has demonstrated its distinctive characteristics and representation in the field of agricultural development. The research area is primarily devoted to the cultivation of corn, wheat, rice, and other crops. Maize, in particular, was planted on 3864.37 thousand hectares in Henan Province in 2023, underscoring its pivotal role in the local agricultural economy. Nanzhang Village, situated in Shanyang District, Jiaozuo City, is located within the Huang-Huai-Hai Plain, which is classified as belonging to the temperate monsoon climate zone. This climatic zone is characterized by four distinct seasons, sufficient light and abundant heat, which provides an ideal environment for the growth of young maize.
In this study, model training and optimization were conducted in research area A, while experiments on model generalization application were carried out in research area B. The distribution of Study Area A and Study Area B is shown in Figure 1, where the area of Study Area A is 6545 square meters, the maize varieties planted are for Xingmin 176 with self-retaining loose seed, and the maize sowing time is 8 June 2023. Xingmin 176 is a variety that has demonstrated positive performance in Henan Province and has high yield potential based on historical data. Its characteristics include purple leaf sheaths at the seedling stage, green leaf blades, vigorous growth, and a more compact plant shape. Establishing a training and validation dataset based on these characteristics helps to provide more reliable results for practical applications. Research area B covers an area of 5445 square meters and is planted with the Fengshou 2 maize variety, which was sown on 6 June 2023. Selecting different maize varieties for this study enables a deeper understanding of the influence of segmentation algorithm accuracy across various instances, thereby facilitating the optimization of algorithm parameters and model design.

2.1. Data Acquisition

The dataset for this study was collected on 4 July 2023, in the eastern cornfield of Nanzhang Village, Shanyang District, Jiaozuo City, Henan Province. Data were collected using a DJI Phantom 4 Multispectral ( DJI-Innovations Inc., Shenzhen, China), with the flight altitude set at 8.5 m, a side-by-side overlap of 65%, and a heading overlap of 65%. The weather on that day was clear and calm, thereby eliminating the possibility of image distortion due to weather conditions, and ensuring an image resolution of 0.4 cm. A total of 2238 aerial photographs were obtained.

2.2. Data Preprocessing

After ensuring that the image quality of the UAV meets the research requirements, this study used version 2.1 Sablefish of the open source software OpenMVG [38] for Structure from Motion (SfM) processing to reconstruct the multispectral ortho image. By analyzing images from different viewing angles, SfM technology can reconstruct the three-dimensional structure of the scene without additional measurement equipment. OpenMVG is used to extract feature points from the images and calculate the camera pose to generate sparse point cloud data. Based on these data, we construct a Digital Elevation Model (DEM) that describes the elevation information of the surface. Combined with camera pose and automatic calibration parameters, we further compute the orthophoto image of the multispectral band. Finally, orthophoto images of RGB (Red–Green–Blue), NRG (Near Infrared–Red–Green), NER (Near Infrared–Red Edge–Red) and other bands are synthesized, and the results are shown in Figure 2:

2.3. Dataset Construction

In order to ensure the effectiveness of detection and segmentation of small targets, we segmented the orthophoto into 640 × 640-pixel image slices. To avoid missed detections caused by targets located between the slices, we set the overlap rate of the slices to 10%. We utilized the SAM [37] for the semi-automatic labeling of individual maize plants. Compared with the weakly supervised learning method [32], this approach achieves a more accurate delineation of monocot maize through manual intervention. Compared to the traditional manual labeling method, the efficiency of semi-automatic labeling using the SAM-based approach has significantly improved, reducing the average labeling time per maize plant to less than 6 s. As illustrated in Table 1, the reduction in annotation costs further demonstrates the advantages of SAM-based semi-automatic annotation in reducing both time and labor inputs for dataset annotation. As depicted in Figure 3, we compare the traditional manual annotation method with the SAM-based semi-automatic annotation method, where the green plus sign (‘+’) on the red contour line indicates the key turning points in the annotation process. The results indicate that the masks generated by the SAM-based semi-automatic labeling method have finer edges and exhibit higher accuracy compared to those produced by the traditional manual labeling method. The labeling results of some images are shown in Figure 4, where 488 images (80%) from Study Area A were used as the training sample set, and 123 images (20%) were used as the validation sample set. In addition, an independent test set consisted of 723 images from Study Area B. This test set was only used to test the generalization ability of the model and did not participate in the training process of the model. According to the statistical results, a total of 24,064 maize plants were labeled in Study Area A, and 29,226 maize plants were labeled in Study Area B. In order to quantify the number of instances in the training, validation, and testing datasets, we used the COCO dataset’s size classification criterion for targets. According to this criterion, small targets refer to those instances smaller than 32 × 32 pixels, medium targets represent instances between 32 × 32 and 96 × 96 pixels, and large targets refer to those instances larger than 96 × 96 pixels. The specific statistics of the number of instances are shown in Table 2.

2.4. The SAM Model

The Segment Anything Model (SAM) [37] is a segmentation model released by Meta in April 2023. It provides the latest results in zero-sample segmentation and is trained on a large-scale SA-1B dataset containing more than 11 million images and 1 billion markers. SAM is capable of segmenting any image or object in any video, eliminating the need for additional training information. The algorithm is utilized not only for satellite image segmentation [39] but also for diverse image segmentation tasks, including medical image segmentation [40], livestock tracking mask extraction [41], and autonomous robot frameworks [42].
In this study, the SAM model is employed to perform the rapid, semi-automatic labeling of point cues for young maize in the image, and the labeling process is depicted in Figure 5. The model consists of three components: an image encoder, a prompt encoder, and a mask decoder.

2.4.1. Image Encoder

The image encoder is structured as a Vision Transformer (ViT) [43]-based masked autoencoder (MAE) [44] for extracting image masks from high-resolution image inputs, which are fed into a mask decoder along with cued embeddings that have been feature extracted.

2.4.2. Prompt Encoder

The input prompts of SAM are categorized into two types: one for sparse prompts, such as dots, bounding boxes, and text; and another for dense prompts, such as masks. The sparse cues are processed by the cue encoder to generate cue embeddings; point cues and bounding box cues are transformed through positional encoding, whereas text cues undergo processing by the text encoder in Contrastive Language-Image Pre-Training (CLIP) [45]. Mask cues undergo convolution and are subsequently summed element by element with the image features. In this study, we utilize point cues to activate the SAM’s zero-sample capability. Given a point cue, SAM can automatically segment a corresponding object from an image within seconds. If the mask lacks precision, it can be refined by adding cues to either missing or excessive regions. This mask can be readily converted into an annotation. This process can be regarded as a semi-automatic labeling method for dataset annotation.

2.4.3. Mask Decoder

The mask decoder efficiently converts image features, cue embeddings, and class tokens into masks. This is achieved using the SAM mask decoder, which utilizes a lightweight Transformer [46] decoder block to convert class tokens (young maize seedlings), cue embeddings, and image features into a dynamic classifier. Subsequently, it computes the probability of each pixel in the image to produce a mask.

2.5. The YOLOv8 Model

Compared to its predecessors, the YOLOv8 model boasts enhanced speed and accuracy. YOLOv8 is categorized into five distinct model sizes based on network depth and feature map dimensions: YOLOv8n, YOLOv8s, YOLOv8m, YOLOv8l, and YOLOv8x. YOLOv8n (YOLOv8 Nano) is the fastest and smallest, while YOLOv8x (YOLOv8 Extra Large) is the most accurate but slowest among them. According to the network architecture diagram, YOLOv8 can be divided into four parts: Input, Backbone, Neck, and Head. The network architecture of YOLOv8-Seg is illustrated in Figure 6:

2.5.1. Input

The primary objective is to resize images to the required training size and perform data augmentation operations. YOLOv8 utilizes an adaptive image scaling method to resize images adaptively, minimizing the width of black borders to reduce computational overhead in image processing. Data augmentation is a technique for generating additional equally effective data based on a limited dataset without altering the essential information in the images. This approach enhances the model’s generalization ability by compelling it to learn more resilient features. Common data augmentation techniques comprise horizontal/vertical flipping, random cropping, scaling, and rotation, among others. During the training of this study, the Mosaic [47] online data augmentation method was disabled in the final 10 training epochs. The Mosaic data augmentation technique involves randomly cropping and resizing four images, which are then spliced together into a single image for use in training data.

2.5.2. Backbone Module

The backbone network is used for extracting target features, and the YOLOv8 backbone consists of the CBS convolution module, C2f module, and Spatial Pyramid Pooling Fast (SPPF) module.
(A)
CBS convolutional module
The CBS convolution module consists of convolutional layers, batch normalization layers (BN), and the SiLU activation function, as illustrated in Figure 7.
The convolution operation involves traversing the image from the top left to the bottom right with a fixed stride, multiplying the pixel values covered by the fixed-size convolutional kernel at each position, and summing them up for computation. The result of this process is the feature map matrix. Common convolutional kernel sizes include 1 × 1 and 3 × 3, among others. The 3 × 3 convolution is used to expand the receptive field, while the 1 × 1 convolutional kernel is used to reduce the number of parameters. It is worth noting that the weight values of the convolutional kernel are shared across the entire image.
During model training, parameter updates can lead to changes in the output distribution of subsequent layers, thereby affecting the speed of model training. To address this issue, batch normalization techniques normalize the input data by mean and variance, ensuring a fixed distribution for each layer’s input data. This accelerates the gradient convergence speed of the model while preventing issues such as gradient vanishing and exploding.
The activation function used in the YOLOv8 model is the SiLU function, which is a special case of the Swish function [48] expressed as shown in Equation (1).
S w i s h x = x 1 + e x p ( β x )
(B)
C2f module
The C2f module is designed based on the C3 module of YOLOv5 and the concept of ELAN, controlling the gradient path length to achieve more effective learning and convergence. This allows the YOLOv8 model to obtain richer gradient information while maintaining its lightweight nature. The structure of the C2f module is illustrated in Figure 8.
(C)
SPPF module
The SPPF module transforms the parallel structure of the max-pooling layers into a serial structure and standardizes the convolutional layer size to 1 × 1. Experimental results demonstrate that this improvement further optimizes performance. The structure of the SPPF module is illustrated in Figure 9.

2.5.3. Neck Module

The Neck section employs the Path Aggregation Feature Pyramid Network (PAFPN) [49] structure for multiscale feature fusion, aiming to integrate features from different levels to extract richer information.

2.5.4. Head Module

YOLOv8 is a top–down instance segmentation model based on object detection, where the Head section decomposes the instance segmentation task into two parallel branch tasks. The image segmentation branch utilizes a fully convolutional network to generate a set of prototype masks, while the object detection branch performs bounding box prediction and class prediction. Additionally, each predicted box is encoded with a set of mask coefficients to represent instances. Subsequently, the two branches are combined through linear combination calculation to output instance masks, generating prototype masks as illustrated in Figure 10.

2.6. Evaluation Indicators

This study aims to achieve the localization and segmentation of each young maize plant with the ultimate goal of addressing practical challenges. Consequently, this study utilized the mean average precision (mAP) as the accuracy evaluation metric. The mAP50 (IoU of 50) and mAP50-95 (average precision at IoU ranging from 0.5 to 0.95 with a step size of 0.05), served as primary indicators for assessing model accuracy.
The precision rate (P) denotes the probability that a predicted positive example is actually a positive sample, as shown in Equation (2).
P = T P T P + F P
where TP is the number of correctly detected maize samples and FP is the number of incorrectly detected maize samples.
Recall (R) is a metric of model detection generalization, expressed as the proportion of correctly detected labels to the total true labels, as shown in Equation (3).
R = T P T P + F N
where FN is the number of undetected maize samples.
Mean average precision (mAP) represents the average of the network’s AP values for all categories in the dataset. The specific calculation expression is shown in Equation (4).
m A P = 1 m A P ( i )
where m is the number of labeled categories, AP is the area under the P-R curve, which responds to how well the model recognizes a category, and the P-R curve is the curve drawn based on the P and R values of the category.

2.7. Experimental Environment

The experimental platform for this study employs the Windows 10 operating system, an Intel (R) Core (TM) i9-9900K CPU processor with a main frequency of 3.60 GHz, and a GeForce RTX 2080 Ti GPU with 11 GB of video memory. The model training framework used is Pytorch 2.0.1, with specific training parameters detailed in Table 3:

3. Experiments and Analysis of Results

3.1. Comparison and Analysis of Different Instance Segmentation Models

To verify the effectiveness of the YOLOv8 model, three distinct band combinations were selected for experimentation: RGB (Red, Green, Blue), NER (Near-Infrared, Red Edge, Red), and NRG (Near Infrared, Red, Green). In the experiments, we compared six instance segmentation network models including YOLOv8, Mask R-CNN [50], Mask Scoring R-CNN [51], PointRend [52], and YOLOv5 [53] models. The experimental results are shown in Table 4, and there are some differences in different band combinations in different algorithmic models. Overall, the NRG and NER bands exhibit similar accuracy across all models. However, in the Cascade Mask R-CNN model, the NRG band outperforms the NER band significantly. Particularly in the YOLOv8 model, the NRG band achieves the highest accuracy among all bands and across all models. In contrast, the RGB band’s performance was generally lower than that of the other two bands across all models. This discrepancy may arise from the varying sensitivities of the bands to feature characteristics coupled with the disparate capabilities of algorithmic models to process band information.
In this experimental study, the YOLOv8 model exhibits excellent accuracy, which is attributed to its advanced deep learning one-shot detection framework. This framework processes images rapidly, accurately identifying young maize seedling plants. The YOLOv5 model also uses the one-shot detection framework, but the YOLOv8 model uses a C2f structure with a richer gradient flow and adjusts the number of channels differently for different scales of the model. The Head section of the YOLOv8 model transitioned from a coupled-head to a decoupled head structure, thereby segregating the classification and detection heads, and shifting from Anchor-Based to Anchor-Free. These changes may be the reason why the YOLOv5 model is ranked second. The PointRend model achieves more accurate segmentation through pixel-level rendering technology, enabling the model to process the image at a fine-grained level and improve segmentation accuracy. In contrast, the Mask R-CNN and Mask Scoring R-CNN detection frameworks rely on anchors and result in higher computational complexity. This may affect their segmentation accuracy when dealing with complex scenes. The Cascade Mask R-CNN model, a multi-stage detection model, has its final outcome influenced by the performance at each stage. Although this model may exhibit high accuracy in some cases, it is computationally expensive and may be less applicable in resource-limited environments. Furthermore, the multi-stage detection process means the final outcome is contingent upon each stage’s performance, potentially explaining lower accuracy levels.
In summary, both the band combination and model design greatly influence the performance of instance segmentation in remote sensing images. Future research could focus on optimizing the model structure to enhance both the accuracy and efficiency of instance segmentation.

3.2. Analysis of Overcrowded Seedling Plants

During maize cultivation, planting density, an important agronomic parameter, directly affects yield and quality [54,55]. In this study, we employed the YOLOv8m model to assess the planting density of maize in Study Area B and analyzed the relationship between the detected planting density and the actual planting density, as illustrated in Figure 11. The variation in color depth within the figure represents the superposition effect of planting density; darker areas suggest that multiple locations share the same planting density, leading to an overlap of points and deeper colors. The results demonstrate that the YOLOv8m model exhibits high detection accuracy and reliability.
Based on the model’s strong performance, we further used the YOLOv8m model for the study of overcrowded zones. Overcrowding may lead to various issues, including heightened competition for light, nutrients, and water among plants, an increased risk of pests and diseases, and reduced efficiency of field operations. Thus, the identification and analysis of overcrowded zones in maize is imperative. The NRG-band image, which performs better in the YOLOv8 model, was chosen for the experiment to achieve the precise instance segmentation of young maize seedling plants, aiming to obtain the precise location information of each maize plant. By calculating the coordinates of the center point of each maize plant and taking into account the spacing criteria recommended by the maize field management program, we formulated a corresponding numerical specification to determine the circle’s size. Per the field management program, maize plants should be spaced between 25 and 30 cm apart. Given the spatial resolution of the remote sensing data utilized, we calculated the circle’s radius to be 62.5 pixel units, serving as the foundation for mapping the circular area centered on each maize plant’s center point. Additionally, we evaluated the probability of other plants’ presence or absence within each circular area to gauge planting density. To identify regions of high planting density, we set a probability threshold above which areas were considered densely planted. To visually represent areas of high planting density, we utilized cross-filament markers to denote them. In the future, this technology is anticipated to be deployed on drones or other monitoring devices for the real-time monitoring of maize plantations.
The specific effect is illustrated in Figure 12. The image depicts the result of the YOLOv8m model recognition using NRG remote sensing image, and it illustrates the positional relationship between the real coordinates and the predicted coordinates. The white outline lines represent individual maize seedling plants identified by the model, while the areas indicated by the cross-filaments highlight overdense areas identified by the model. These zones were automatically defined by the model based on the density of the seedling plants, emphasizing areas where field management might be required to optimize plant growth conditions.
By carefully comparing the recognition results, we can observe that some recognition results have some deviations from the real situation, and these deviations are mainly reflected in the recognition accuracy of the model for maize seedling plants. Sometimes, the model failed to accurately recognize independent maize seedling plants but misjudged the seedling plants with closer growth as a single plant, which led to the misreporting of overcrowded areas. This phenomenon may be caused by the model’s algorithm’s lack of accuracy in handling subtle densification differences between plants.
Despite these biases, the YOLOv8 model has demonstrated overall utility in recognizing the density of maize seedling plants. However, to enhance recognition accuracy, future research should consider developing a more comprehensive database and refining the model to better adapt to complex field conditions. This approach will enable agricultural workers to make more informed field management decisions, optimize crop growth conditions, and enhance yields.

4. Discussion

This section outlines three experiments conducted to explore the factors influencing the segmentation accuracy of maize instances. First, we examined the effect of varying parameter counts in the YOLOv8 model on segmentation accuracy. Second, experiments were conducted to assess the generalization capability of various mainstream instance segmentation models in Study Area B. Finally, we explored the performance of different spatial resolutions in the YOLOv8 model. The results and analysis of these experiments provide critical insights and guidance for advancing maize instance segmentation technology.

4.1. Effect of Different Parametric Quantities on Segmentation Accuracy

Five model configurations (n, s, m, l, and x) and three band combination methods (NRG, NER, and RGB) of YOLOv8 underwent training and validation. The results are presented in Table 5, where the input image sizes were consistently 640 × 640. This illustrates that regardless of the band combination, the mean average precision (mAP) demonstrates an increasing trend with the model’s complexity. However, with the model’s increasing complexity, there is also an increase in detection time, the number of floating point operations (GFLOPs), and the parameter count. Using the NRG band as an example, in comparison to the YOLOv8n model, the subsequent four models (s, m, l, and x) exhibited improvements of 2.8%, 4.4%, 5.7%, and 6.1% in object detection and 2.8%, 4.2%, 4.6%, and 4.8% in semantic segmentation, respectively. Notably, the YOLOv8m and YOLOv8x models perform similarly in detection and segmentation with a significant difference in FLOPs at 234.1 G and in parameter size at 44.512 MB. Given the YOLOv8m model’s strong performance in inference time, detection accuracy, and parameter count, it is recommended as a foundational model for ongoing research applications.
Second, the table reveals that the accuracy of the RGB band is notably lower than that of the NER and NRG bands. For example, with the YOLOv8m model, the mAPBox50-95 and mAPSeg50-95 of the NRG band reached 0.794 and 0.618, respectively. In contrast, although the inference speed increased by 0.8 ms, the detection and segmentation accuracy of the NER band saw a minor decline. Although few studies have directly elucidated the superiority of NRG bands across all relevant domains, existing research and remote sensing theories offer indirect support for this perspective. Given the high sensitivity of NRG bands to vegetation [56,57,58,59,60,61] and soil [62,63,64], their broad application in various areas [65,66,67,68,69,70,71], and utility in calculating vegetation indices [72,73,74,75,76,77,78,79,80], selecting NRG bands is a reasonable and effective strategy for routine studies in maize fields. Future studies will employ NRG bands as the primary bands to deliver reliable solutions for agricultural applications.

4.2. Generalization Experiment

The dataset constructed in Study Area B was utilized for generalization experiments. Six mainstream instance segmentation models, including YOLOv8, were chosen for the experiments, and the results are depicted in Table 6. These results unequivocally indicate that the YOLOv8 model presents considerable advantages over other algorithms. Although YOLOv8 possesses a greater number of parameters and higher computational requirements compared to the YOLOv5 model, its detection and segmentation accuracies (50-95) saw improvements of 3.1% and 3.4%, respectively, for the same model size (M model). This emphasizes the superiority of YOLOv8 in terms of overall performance, stability, and reliability, demonstrating its greater adaptability and generalizability. The YOLOv8 model is likely to offer greater advantages in various complex scenarios and tasks.
In addition, the inference time per image is decreased by 30.3 milliseconds relative to the PointRend model, despite the higher FLOPs of the YOLOv8m model. Detection accuracy and segmentation accuracy (50–95) have seen improvements of 15.7% and 11.4%, respectively. These results thoroughly validate the YOLOv8m model’s excellent performance in detection and segmentation tasks.
To more effectively illustrate the performance of different instance segmentation methods and visualize the detection results of each model, we selected three representative scenarios from the dataset for illustration: a scenario with relatively short, sparse maize mixed with weeds (Figure 13); a scenario with densely growing maize (Figure 14); and a scenario with normally growing maize mixed with other plants (Figure 15). These three scenarios encompass various conditions encountered during maize growth and facilitate a comprehensive evaluation of the performance of different instance segmentation models.
In this scenario, where the maize is relatively short, sparse, and mixed with weeds, the challenge lies in the model’s ability to distinguish between maize plants and weeds while accurately segmenting each instance despite the sparse distribution of plants. As Figure 13 illustrated in the figure, the PointRend, Mask Scoring R-CNN, and Mask R-CNN models exhibit misdetection by erroneously identifying other plants in the upper right corner as maize. In contrast, the YOLOv8 model demonstrates superior performance in recognizing and segmenting thin and elongated leaves of maize. However, all models exhibit a degree of under-detection for maize plants adjacent to weeds, especially for those that are short in height.
In the case of dense maize growth, the leaves overlap heavily, and the model must be able to handle this complexity and segment each instance accurately. For this reason, we deliberately chose the case of dense maize plants for demonstration. As can be seen in Figure 14, with the exception of the YOLOv8 model, all other models have the situation of recognizing two maize plants as one and fail to segment the maize leaves accurately and finely.
In the case where the maize grows normally and is mixed with other plants, the model must be able to effectively discriminate between different types of plants and accurately segment them into instances. The results, depicted in Figure 15, reveal that even the YOLOv5 model exhibits mixed detection omissions in this scenario, while the Cascade Mask R-CNN model shows substantial areas of omission. Particularly in scenes featuring a significant mixture of other plants in the middle, all models exhibit some degree of missed detection. Nevertheless, as evident from the figure, YOLOv8 exhibits the lowest leakage rate and achieves fine segmentation of the recognized corn plants’ leaves.
The YOLOv8 model demonstrated robust performance and generalization capabilities despite the variation in maize varieties between Study Area A and Study Area B. The YOLOv8 model excelled in detecting and generalizing across maize varieties. Its superiority over other algorithms lies in its ability to effectively adapt to different maize varieties while maintaining high detection and segmentation accuracy. Consequently, it is concluded that the YOLOv8 model is exceptionally adaptable and robust, handling maize varieties with varying characteristics and offering a reliable solution for agricultural applications. It is worth noting that although the YOLOv8 model performed best in the experiments, there is still room for improvement in the accuracy of segmenting the boundaries of a single maize plant especially when dealing with elongated leaves and mixed weeds. This finding offers valuable insights for future research directions and innovations.

4.3. Model Performance at Different Spatial Resolutions

The initial data for this study were derived from aerial images acquired by a UAV flying at an altitude of 8.5 m with a ground spatial resolution (GSD) of 0.4 cm/pixel. A primary advantage of this altitude is its capability to capture high-resolution images, thus offering clearer and more detailed data to support the instance segmentation of young maize. These high-precision data facilitate a more accurate identification and segmentation of plant instars during data annotation, thereby ensuring the reliability and accuracy of the study results. However, we recognize that such low flight altitudes may pose practical challenges in managing large areas. To identify the optimal imaging altitude for segmenting young maize instars and enhance the detection performance in real-world applications, we adjusted the spatial resolution to explore the relationship between spatial resolution and instance segmentation accuracy. The YOLOv8m model served as the base model, with the NRG band—demonstrating superior performance in previous experiments—selected for further experimental analysis, as detailed in Table 7.
To more clearly illustrate the relationship between spatial resolution and instance segmentation accuracy, mAP50 is utilized as an indicator, and a line graph depicting the change in instance segmentation accuracy at various spatial resolutions is presented in Figure 16. The figure clearly shows that both detection accuracy and segmentation accuracy decrease with diminishing spatial resolution. As spatial resolution declines from 1.333 to 1.600 cm/pixel, there is a significant drop in accuracy—the mAPBox50 decreased by 1.3% and the mAPSeg50 decreased by 3.3%.
At a GSD of 1.333 cm/pixel, mAPBox50 achieves 91.4%, and mAPSeg50 achieves 89.7%, fulfilling the accuracy requirements for routine applications. Per Equation (5), with a GSD of 1.333 cm/pixel, the corresponding height (H) is 28.326 m, meeting the demands of large-scale daily scenarios for monitoring and managing young maize growth.
H = f × G S D a

5. Conclusions

To achieve the precise positioning and segmentation of young maize seedlings in daily large-area scenarios, facilitating the monitoring and management of maize growth conditions, this study focuses on employing the YOLOv8 model for maize instance segmentation in UAV-based remote sensing. We systematically explore the impact of different multispectral band combinations, spatial resolutions, and model parameters on segmentation accuracy. The main conclusions are as follows:
(1) Combined with the SAM image segmentation large-scale model, it can realize the efficient annotation of young maize seedling plants. This approach facilitates the creation of a representative maize seedling plant instance segmentation dataset, addressing current research challenges like annotation data inaccuracy and training data limitations.
(2) The YOLOv8 model shows excellent segmentation accuracy under different multispectral data combinations and different resolutions, especially in NRG synthetic data; the accuracy of mAPBox50-90 reaches 0.794 and that of mAPSeg50-95 reaches 0.618. In the generalization experiments, YOLOv8 also performs well, which proves its applicability and stability under different environments and conditions.
(3) Spatial resolution has a significant effect on the segmentation accuracy of young maize instances at a given flight altitude. Appropriate resolution can meet the needs of practical applications, which in turn improves the cost-effectiveness of applications in real-world scenarios. This finding is of great practical importance for monitoring and managing the growth of young maize in everyday large area scenarios.
This study offers a crucial theoretical foundation and practical guidance for deploying UAV remote sensing technology in agricultural automation and intelligent monitoring. In particular, the excellent performance of the YOLOv8 model in the generalization experiment further proves its great potential in the field of precision agriculture. Subsequent research will be devoted to optimizing the network structure to improve the segmentation detection ability of the model by adding the attention mechanism and changing the loss function. Meanwhile, we will also explore methods such as model pruning and knowledge distillation to reduce the number of parameters in order to improve the usability of the model in the case of limited computational power. Through these methods, we expect to further improve the accuracy and efficiency of segmentation of young maize seedling plant instances and provide effective technical support for precision agriculture.

Author Contributions

Conceptualization, T.G. and H.Y.; methodology, T.G. and H.Y; software, P.L., X.Y. and R.M; validation, T.G.; writing—original draft preparation, T.G.; writing—review and editing, H.Y., X.Y. and R.M.; visualization, T.G.; supervision, H.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Natural Science Foundation of China (U1304402, 41977284) and Laboratory of Mine Spatio-Temporal Information and Ecological Restoration, MNR (No. KLM202310).

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available due to privacy.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Shu, M.Y.; Shen, M.Y.; Zuo, J.Y.; Yin, P.F.; Wang, M.; Xie, Z.W.; Tang, J.H.; Wang, R.L.; Li, B.G.; Yang, X.; et al. The Application of UAV-Based Hyperspectral Imaging to Estimate Crop Traits in Maize Inbred Lines. Plant Phenomics 2021, 2021, 9890745. [Google Scholar] [CrossRef] [PubMed]
  2. Shi, M.; Zhang, S.; Lu, H.; Zhao, X.; Wang, X.; Cao, Z. Phenotyping Multiple Maize Ear Traits from a Single Image: Kernels per Ear, Rows per Ear, and Kernels per Row. Comput. Electron. Agric. 2022, 193, 106681. [Google Scholar] [CrossRef]
  3. Yue, J.; Guo, W.; Yang, G.; Zhou, C.; Feng, H.; Qiao, H. Method for Accurate Multi-Growth-Stage Estimation of Fractional Vegetation Cover Using Unmanned Aerial Vehicle Remote Sensing. Plant Methods 2021, 17, 51. [Google Scholar] [CrossRef] [PubMed]
  4. Yue, J.; Tian, Q.; Liu, Y.; Fu, Y.; Tian, J.; Zhou, C.; Feng, H.; Yang, G. Mapping Cropland Rice Residue Cover Using a Radiative Transfer Model and Deep Learning. Comput. Electron. Agric. 2023, 215, 108421. [Google Scholar] [CrossRef]
  5. Zhuang, L.; Wang, C.; Hao, H.; Li, J.; Xu, L.; Liu, S.; Guo, X. Maize Emergence Rate and Leaf Emergence Speed Estimation via Image Detection under Field Rail-Based Phenotyping Platform. Comput. Electron. Agric. 2024, 220, 108838. [Google Scholar] [CrossRef]
  6. Goggin, F.L.; Lorence, A.; Topp, C.N. Applying High-Throughput Phenotyping to Plant–Insect Interactions: Picturing More Resistant Crops. Curr. Opin. Insect Sci. 2015, 9, 69–76. [Google Scholar] [CrossRef]
  7. Varshney, R.K.; Thudi, M.; Pandey, M.K.; Tardieu, F.; Ojiewo, C.; Vadez, V.; Whitbread, A.M.; Siddique, K.H.M.; Nguyen, H.T.; Carberry, P.S.; et al. Accelerating Genetic Gains in Legumes for the Development of Prosperous Smallholder Agriculture: Integrating Genomics, Phenotyping, Systems Modelling and Agronomy. J. Exp. Bot. 2018, 69, 3293–3312. [Google Scholar] [CrossRef]
  8. Zhao, C.; Zhang, Y.; Du, J.; Guo, X.; Wen, W.; Gu, S.; Wang, J.; Fan, J. Crop Phenomics: Current Status and Perspectives. Front. Plant Sci. 2019, 10, 714. [Google Scholar] [CrossRef]
  9. Tang, L.; Shao, G. Drone Remote Sensing for Forestry Research and Practices. J. For. Res. 2015, 26, 791–797. [Google Scholar] [CrossRef]
  10. Baresel, J.P.; Rischbeck, P.; Hu, Y.; Kipp, S.; Hu, Y.; Barmeier, G.; Mistele, B.; Schmidhalter, U. Use of a Digital Camera as Alternative Method for Non-Destructive Detection of the Leaf Chlorophyll Content and the Nitrogen Nutrition Status in Wheat. Comput. Electron. Agric. 2017, 140, 25–33. [Google Scholar] [CrossRef]
  11. Zhou, L.; Gu, X.; Cheng, S.; Yang, G.; Shu, M.; Sun, Q. Analysis of Plant Height Changes of Lodged Maize Using UAV-LiDAR Data. Agriculture 2020, 10, 146. [Google Scholar] [CrossRef]
  12. Zhou, X.; Zheng, H.B.; Xu, X.Q.; He, J.Y.; Ge, X.K.; Yao, X.; Cheng, T.; Zhu, Y.; Cao, W.X.; Tian, Y.C. Predicting Grain Yield in Rice Using Multi-Temporal Vegetation Indices from UAV-Based Multispectral and Digital Imagery. ISPRS J. Photogramm. Remote Sens. 2017, 130, 246–255. [Google Scholar] [CrossRef]
  13. Potgieter, A.B.; George-Jaeggli, B.; Chapman, S.C.; Laws, K.; Suárez Cadavid, L.A.; Wixted, J.; Watson, J.; Eldridge, M.; Jordan, D.R.; Hammer, G.L. Multi-Spectral Imaging from an Unmanned Aerial Vehicle Enables the Assessment of Seasonal Leaf Area Dynamics of Sorghum Breeding Lines. Front. Plant Sci. 2017, 8, 1532. [Google Scholar] [CrossRef]
  14. Song, P.; Wang, J.; Guo, X.; Yang, W.; Zhao, C. High-Throughput Phenotyping: Breaking through the Bottleneck in Future Crop Breeding. Crop J. 2021, 9, 633–645. [Google Scholar] [CrossRef]
  15. Leaf Area Index Estimation of Pergola-Trained Vineyards in Arid Regions Using Classical and Deep Learning Methods Based on UAV-Based RGB Images—ScienceDirect. Available online: https://www.sciencedirect.com/science/article/pii/S0168169923001114 (accessed on 1 April 2024).
  16. Zhang, J.; Cheng, T.; Guo, W.; Xu, X.; Qiao, H.; Xie, Y.; Ma, X. Leaf Area Index Estimation Model for UAV Image Hyperspectral Data Based on Wavelength Variable Selection and Machine Learning Methods. Plant Methods 2021, 17, 49. [Google Scholar] [CrossRef] [PubMed]
  17. Pei, S.; Zeng, H.; Dai, Y.; Bai, W.; Fan, J. Nitrogen Nutrition Diagnosis for Cotton under Mulched Drip Irrigation Using Unmanned Aerial Vehicle Multispectral Images. J. Integr. Agric. 2023, 22, 2536–2552. [Google Scholar] [CrossRef]
  18. Hu, J.; Feng, H.; Wang, Q.; Shen, J.; Wang, J.; Liu, Y.; Feng, H.; Yang, H.; Guo, W.; Qiao, H.; et al. Pretrained Deep Learning Networks and Multispectral Imagery Enhance Maize LCC, FVC, and Maturity Estimation. Remote Sens. 2024, 16, 784. [Google Scholar] [CrossRef]
  19. Han, L.; Yang, G.; Dai, H.; Xu, B.; Yang, H.; Feng, H.; Li, Z.; Yang, X. Modeling Maize Above-Ground Biomass Based on Machine Learning Approaches Using UAV Remote-Sensing Data. Plant Methods 2019, 15, 10. [Google Scholar] [CrossRef] [PubMed]
  20. Miao, T.; Zhu, C.; Xu, T.; Yang, T.; Li, N.; Zhou, Y.; Deng, H. Automatic Stem-Leaf Segmentation of Maize Shoots Using Three-Dimensional Point Cloud. Comput. Electron. Agric. 2021, 187, 106310. [Google Scholar] [CrossRef]
  21. Turgut, K.; Dutağacı, H.; Rousseau, D. RoseSegNet: An Attention-Based Deep Learning Architecture for Organ Segmentation of Plants. Biosyst. Eng. 2022, 221, 138–153. [Google Scholar] [CrossRef]
  22. Lu, Z.; Qi, L.; Zhang, H.; Wan, J.; Zhou, J. Image Segmentation of UAV Fruit Tree Canopy in a Natural Illumination Environment. Agriculture 2022, 12, 1039. [Google Scholar] [CrossRef]
  23. Ren, M.; Zemel, R.S. End-to-End Instance Segmentation with Recurrent Attention. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 293–301. [Google Scholar] [CrossRef]
  24. Romera-Paredes, B.; Torr, P.H.S. Recurrent Instance Segmentation. In Proceedings of the Computer Vision—ECCV 2016, Amsterdam, The Netherlands, 11–14 October 2016; Leibe, B., Matas, J., Sebe, N., Welling, M., Eds.; Springer International Publishing: Cham, Switzerland, 2016; pp. 312–329. [Google Scholar]
  25. Scharr, H.; Minervini, M.; Fischbach, A.; Tsaftaris, S. Annotated Image Datasets of Rosette Plants. In Proceedings of the European Conference on Computer Vision, Zurich, Switzerland, 6–12 September 2014. [Google Scholar]
  26. Yin, X.; Liu, X.; Chen, J.; Kramer, D.M. Multi-Leaf Tracking from Fluorescence Plant Videos. In Proceedings of the 2014 IEEE International Conference on Image Processing (ICIP), Paris, France, 27–30 October 2014; pp. 408–412. [Google Scholar]
  27. Zhang, J.; Wang, X.; Liu, J.Y.; Zhang, D.; Lu, Y.; Zhou, Y.H.; Sun, L.; Hou, S.L.; Fan, X.F.; Shen, S.; et al. Multispectral Drone Imagery and SRGAN for Rapid Phenotypic Mapping of Individual Chinese Cabbage Plants. Plant Phenomics 2022, 2022, 0007. [Google Scholar] [CrossRef] [PubMed]
  28. Bai, Y.; Shi, L.; Zha, Y.; Liu, S.; Nie, C.; Xu, H.; Yang, H.; Shao, M.; Yu, X.; Cheng, M.; et al. Estimating Leaf Age of Maize Seedlings Using UAV-Based RGB and Multispectral images. Comput. Electron. Agric. 2023, 215, 108349. [Google Scholar] [CrossRef]
  29. Yang, T.; Zhou, S.; Xu, A.; Ye, J.; Yin, J. An Approach for Plant Leaf Image Segmentation Based on YOLOV8 and the Improved DEEPLABV3+. Plants 2023, 12, 3438. [Google Scholar] [CrossRef] [PubMed]
  30. Qiang, Z.; Shi, J.; Shi, F. Phenotype Tracking of Leafy Greens Based on Weakly Supervised Instance Segmentation and Data Association. Agronomy 2022, 12, 1567. [Google Scholar] [CrossRef]
  31. Bearman, A.; Russakovsky, O.; Ferrari, V.; Fei-Fei, L. What’s the Point: Semantic Segmentation with Point Supervision. In Proceedings of the Computer Vision—ECCV 2016, Amsterdam, The Netherlands, 11–14 October 2016; Leibe, B., Matas, J., Sebe, N., Welling, M., Eds.; Springer International Publishing: Cham, Switzerland, 2016; pp. 549–565. [Google Scholar]
  32. Zhao, Yannan; Deng, Hanbing; Liu, Ting; Zhao, Lulu; Zhao, Kai; Yang, Jing; Zhang, Yufeng A weakly-supervised learning-based method for segmenting maize seedling plant image instances. J. Agric. Eng. 2022, 38, 143–152.
  33. Bawankule, R.; Gaikwad, V.; Kulkarni, I.; Kulkarni, S.; Jadhav, A.; Ranjan, N. Visual Detection of Waste Using YOLOv8. In Proceedings of the 2023 International Conference on Sustainable Computing and Smart Systems (ICSCSS), Coimbatore, India, 14–16 June 2023; pp. 869–873. [Google Scholar]
  34. Hu, Y.; Wang, J.; Wang, X.; Sun, Y.; Yu, H.; Zhang, J. Real-Time Evaluation of the Blending Uniformity of Industrially Produced Gravelly Soil Based on Cond-YOLOv8-seg. J. Ind. Inf. Integr. 2024, 39, 100603. [Google Scholar] [CrossRef]
  35. de Melo Lima, B.P.; de Araújo Barbosa Borges, L.; Hirose, E.; Borges, D.L. A Lightweight and Enhanced Model for Detecting the Neotropical Brown Stink Bug, Euschistus Heros (Hemiptera: Pentatomidae) Based on YOLOv8 for Soybean fields. Ecol. Inform. 2024, 80, 102543. [Google Scholar] [CrossRef]
  36. Tang, Y.; Qian, Y. High-Speed Railway Track Components Inspection Framework Based on YOLOv8 with High-Performance Model deployment. High-Speed Railw. 2024, 2, 42–50. [Google Scholar] [CrossRef]
  37. Kirillov, A.; Mintun, E.; Ravi, N.; Mao, H.; Rolland, C.; Gustafson, L.; Xiao, T.; Whitehead, S.; Berg, A.C.; Lo, W.-Y.; et al. Segment Anything. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 2–3 October 2023. [Google Scholar]
  38. Moulon, P.; Monasse, P.; Marlet, R. Global Fusion of Relative Motions for Robust, Accurate and Scalable Structure from Motion. In Proceedings of the 2013 IEEE International Conference on Computer Vision, Sydney, Australia, 1–8 December 2013; pp. 3248–3255. [Google Scholar]
  39. Zhang, C.; Marfatia, P.; Farhan, H.; Di, L.; Lin, L.; Zhao, H.; Li, H.; Islam, M.D.; Yang, Z. Enhancing USDA NASS Cropland Data Layer with Segment Anything Model. In Proceedings of the 2023 11th International Conference on Agro-Geoinformatics (Agro-Geoinformatics), Novi Sad, Serbia, 15–18 July 2023; pp. 1–5. [Google Scholar]
  40. Pandey, S.; Chen, K.-F.; Dam, E.B. Comprehensive Multimodal Segmentation in Medical Imaging: Combining YOLOv8 with SAM and HQ-SAM Models. In Proceedings of the 2023 IEEE/CVF International Conference on Computer Vision Workshops (ICCVW), Paris, France, 2–6 October 2023; pp. 2584–2590. [Google Scholar]
  41. Noe, S.M.; Zin, T.T.; Tin, P.; Kobyashi, I. Efficient Segment-Anything Model for Automatic Mask Region Extraction in Livestock Monitoring. In Proceedings of the 2023 IEEE 13th International Conference on Consumer Electronics—Berlin (ICCE-Berlin), Berlin, Germany, 2–5 September 2023; pp. 167–171. [Google Scholar]
  42. Dikshit, A.; Bartsch, A.; George, A.; Farimani, A.B. RoboChop: Autonomous Framework for Fruit and Vegetable Chopping Leveraging Foundational Models. arXiv 2023, arXiv:2307.13159. [Google Scholar]
  43. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An Image Is Worth 16x16 Words: Transformers for Image Recognition at Scale. arXiv 2021, arXiv:2010.11929. [Google Scholar]
  44. He, K.; Chen, X.; Xie, S.; Li, Y.; Dollár, P.; Girshick, R. Masked Autoencoders Are Scalable Vision Learners. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 19–25 June 2021. [Google Scholar]
  45. Radford, A.; Kim, J.W.; Hallacy, C.; Ramesh, A.; Goh, G.; Agarwal, S.; Sastry, G.; Askell, A.; Mishkin, P.; Clark, J.; et al. Learning Transferable Visual Models From Natural Language Supervision. In Proceedings of the International Conference on Machine Learning, Vienna, Austria, 18–24 July 2021. [Google Scholar]
  46. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention Is All You Need. Adv. Neural Inf. Process. Syst. arXiv arXiv:1706.03762, 2023.
  47. Bochkovskiy, A.; Wang, C.-Y.; Liao, H.-Y.M. YOLOv4: Optimal Speed and Accuracy of Object Detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
  48. Ramachandran, P.; Zoph, B.; Le, Q.V. Searching for Activation Functions. arXiv 2017, arXiv:1710.05941. [Google Scholar]
  49. Liu, S.; Qi, L.; Qin, H.; Shi, J.; Jia, J. Path Aggregation Network for Instance Segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  50. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask R-CNN. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar]
  51. Huang, Z.; Huang, L.; Gong, Y.; Huang, C.; Wang, X. Mask Scoring R-CNN. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  52. Kirillov, A.; Wu, Y.; He, K.; Girshick, R. PointRend: Image Segmentation As Rendering. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 14–19 June 2020; pp. 9796–9805. [Google Scholar]
  53. Jocher, G.; Chaurasia, A.; Stoken, A.; Borovec, J.; NanoCode012; Kwon, Y.; Michael, K.; Tao, X.; Fang, J.; Imyhxy; et al. Ultralytics/Yolov5: V7.0—YOLOv5 SOTA Realtime Instance Segmentation. Zenodo 2022. [Google Scholar] [CrossRef]
  54. Jin, X.; Liu, S.; Frederic, B.; Hemmerlé, M.; Comar, A. Estimates of Plant Density of Wheat Crops at Emergence from Very Low Altitude UAV Imagery. Remote Sens. Environ. 2017, 198, 105–114. [Google Scholar] [CrossRef]
  55. Liu, S.; Baret, F.; Andrieu, B.; Burger, P.; Hemmerlé, M. Estimation of Wheat Plant Density at Early Stages Using High Resolution Imagery. Front. Plant Sci. 2017, 8, 739. [Google Scholar] [CrossRef] [PubMed]
  56. Jensen, J.R. Remote Sensing of the Environment: An Earth Resource Perspective. In Prentice Hall Series in Geographic Information Science, 2nd ed.; Pearson Prentice Hall: Upper Saddle River, NJ, USA, 2007; ISBN 978-0-13-188950-7. [Google Scholar]
  57. Zhang, Z.; Zhu, L. A Review on Unmanned Aerial Vehicle Remote Sensing: Platforms, Sensors, Data Processing Methods, and Applications. Drones 2023, 7, 398. [Google Scholar] [CrossRef]
  58. Yang, G.; Liu, J.; Zhao, C.; Li, Z.; Huang, Y.; Yu, H.; Xu, B.; Yang, X.; Zhu, D.; Zhang, X.; et al. Unmanned Aerial Vehicle Remote Sensing for Field-Based Crop Phenotyping: Current Status and Perspectives. Front. Plant Sci. 2017, 8, 1111. [Google Scholar] [CrossRef]
  59. BandCombinations.pdf. Available online: https://fsapps.nwcg.gov/gtac/CourseDownloads/Reimbursables/FY20/FHM/Day2/Track3/BandComboCheatSheet.pdf (accessed on 5 June 2024).
  60. Malla, S.; Tuladhar, A.; Quadri, G.J.; Rosen, P. Multi-Spectral Satellite Image Analysis for Feature Identification and Change Detection VAST Challenge 2017: Honorable Mention for Good Facilitation of Single Image Analysis. In Proceedings of the 2017 IEEE Conference on Visual Analytics Science and Technology (VAST), Phoenix, AZ, USA, 3–6 October 2017; pp. 205–206. [Google Scholar]
  61. Langhe, S.; Herbei, M.V.; Sala, F. Use of Remote Sensing Images in Crop Monitoring Case Study: Soybean Crop. Res. J. Agric. Sci. 2020, 52, 53–61. [Google Scholar]
  62. Wigmore, O.; Mark, B.; McKenzie, J.; Baraer, M.; Lautz, L. Sub-Metre Mapping of Surface Soil Moisture in Proglacial Valleys of the Tropical Andes Using a Multispectral Unmanned Aerial Vehicle. Remote Sens. Environ. 2019, 222, 104–118. [Google Scholar] [CrossRef]
  63. Zhang, Y.; Han, W.; Zhang, H.; Niu, X.; Shao, G. Evaluating Soil Moisture Content under Maize Coverage Using UAV Multimodal Data by Machine Learning Algorithms. J. Hydrol. 2023, 617, 129086. [Google Scholar] [CrossRef]
  64. Nguyen, T.T.; Ngo, H.H.; Guo, W.; Chang, S.W.; Nguyen, D.D.; Nguyen, C.T.; Zhang, J.; Liang, S.; Bui, X.T.; Hoang, N.B. A Low-Cost Approach for Soil Moisture Prediction Using Multi-Sensor Data and Machine Learning Algorithm. Sci. Total Environ. 2022, 833, 155066. [Google Scholar] [CrossRef] [PubMed]
  65. Boschetti, M.; Bocchi, S.; Brivio, P.A. Assessment of Pasture Production in the Italian Alps Using Spectrometric and Remote Sensing Information. Agric. Ecosyst. Environ. 2007, 118, 267–272. [Google Scholar] [CrossRef]
  66. Mihai Valentin, H.; Florin, S. Using GIS Technology in Processing and Analyzing Satellite Images—Case Study Cheile Nerei Beusnița National Park, Romania. J. Hortic. For. Biotechnol. 2014, 18, 113–119. [Google Scholar]
  67. Temenos, A.; Temenos, N.; Kaselimi, M.; Doulamis, A.; Doulamis, N. Interpretable Deep Learning Framework for Land Use and Land Cover Classification in Remote Sensing Using SHAP. IEEE Geosci. Remote Sens. Lett. 2023, 20, 1–5. [Google Scholar] [CrossRef]
  68. Dixit, M.; Chaurasia, K.; Kumar Mishra, V. Dilated-ResUnet: A Novel Deep Learning Architecture for Building Extraction from Medium Resolution Multi-Spectral Satellite imagery. Expert Syst. Appl. 2021, 184, 115530. [Google Scholar] [CrossRef]
  69. Hanna, M.M.; Steyn-Ross, D.A.; Steyn-Ross, M. Estimating Biomass for New Zealand Pasture Using Optical Remote Sensing Techniques. Geocarto Int. 1999, 14, 89–94. [Google Scholar] [CrossRef]
  70. Primicerio, J.; Gay, P.; Ricauda Aimonino, D.; Comba, L.; Matese, A.; Di Gennaro, S. NDVI-Based Vigour Maps Production Using Automatic Detection of Vine Rows in Ultra-High Resolution Aerial Images. In Precision Agriculture; Wageningen Academic: Wageningen, The Netherlands, 2015; pp. 465–470. ISBN 978-90-8686-267-2. [Google Scholar]
  71. Yang, C.; Everitt, J.H.; Bradford, J.M.; Escobar, D.E. Escobar Mapping Grain Sorghum Growth and Yield Variations Using Airborne Multispectral Digital Imagery. Trans. ASAE 2000, 43, 1927–1938. [Google Scholar] [CrossRef]
  72. Tucker, C.J. Red and Photographic Infrared Linear Combinations for Monitoring Vegetation. Remote Sens. Environ. 1979, 8, 127–150. [Google Scholar] [CrossRef]
  73. Rouse, J.; Haas, R.H.; Schell, J.A.; Deering, D. Monitoring Vegetation Systems in the Great Plains with ERTS. NASA Spec. Publ. 1973, 351, 309. [Google Scholar]
  74. Qi, J.; Chehbouni, A.; Huete, A.R.; Kerr, Y.H.; Sorooshian, S. A Modified Soil Adjusted Vegetation Index. Remote Sens. Environ. 1994, 48, 119–126. [Google Scholar] [CrossRef]
  75. Rondeaux, G.; Steven, M.; Baret, F. Optimization of Soil-Adjusted Vegetation Indices. Remote Sens. Environ. 1996, 55, 95–107. [Google Scholar] [CrossRef]
  76. Huete, A.R. A Soil-Adjusted Vegetation Index (SAVI). Remote Sens. Environ. 1988, 25, 295–309. [Google Scholar] [CrossRef]
  77. Gitelson, A.A.; Kaufman, Y.J.; Merzlyak, M.N. Use of a Green Channel in Remote Sensing of Global Vegetation from EOS-MODIS. Remote Sens. Environ. 1996, 58, 289–298. [Google Scholar] [CrossRef]
  78. Escadafal, R. Remote Sensing of Arid Soil Surface Color with Landsat Thematic mapper. Adv. Space Res. 1989, 9, 159–163. [Google Scholar] [CrossRef]
  79. Daughtry, C.S.T.; Gallo, K.P.; Goward, S.N.; Prince, S.D.; Kustas, W.P. Spectral Estimates of Absorbed Radiation and Phytomass Production in Corn and Soybean Canopies. Remote Sens. Environ. 1992, 39, 141–152. [Google Scholar] [CrossRef]
  80. Penuelas, J.; Frederic, B.; Filella, I. Semi-Empirical Indices to Assess Carotenoids/Chlorophyll-a Ratio from Leaf Spectral Reflectance. Photosynthetica 1995, 31, 221–230. [Google Scholar]
Figure 1. Overview of the study area.
Figure 1. Overview of the study area.
Plants 13 01842 g001
Figure 2. Orthorectified images of Study Area A and Study Area B.
Figure 2. Orthorectified images of Study Area A and Study Area B.
Plants 13 01842 g002
Figure 3. Comparison between traditional manual annotation and semi-automatic annotation based on SAM model. The green plus sign (‘+’) indicates a key turning point in the annotation process.
Figure 3. Comparison between traditional manual annotation and semi-automatic annotation based on SAM model. The green plus sign (‘+’) indicates a key turning point in the annotation process.
Plants 13 01842 g003
Figure 4. Labeling effect.
Figure 4. Labeling effect.
Plants 13 01842 g004
Figure 5. SAM model drawing labeling flowchart.
Figure 5. SAM model drawing labeling flowchart.
Plants 13 01842 g005
Figure 6. YOLOv8-Seg network structure.
Figure 6. YOLOv8-Seg network structure.
Plants 13 01842 g006
Figure 7. Convolution module.
Figure 7. Convolution module.
Plants 13 01842 g007
Figure 8. C2f module structure.
Figure 8. C2f module structure.
Plants 13 01842 g008
Figure 9. SPPF module.
Figure 9. SPPF module.
Plants 13 01842 g009
Figure 10. Detect Head.
Figure 10. Detect Head.
Plants 13 01842 g010
Figure 11. Comparison of model-detected plant density with real plant density.
Figure 11. Comparison of model-detected plant density with real plant density.
Plants 13 01842 g011
Figure 12. Overcrowding area display. (A) Image edge area; (B) Internal area of image. The green + is the overdense area detected by the model.
Figure 12. Overcrowding area display. (A) Image edge area; (B) Internal area of image. The green + is the overdense area detected by the model.
Plants 13 01842 g012
Figure 13. Comparison of segmentation results of model instances in the scenario of short, sparse and mixed weeds in maize. The red box represents the object detection box. The yellow box represents the area of detail enlargement.
Figure 13. Comparison of segmentation results of model instances in the scenario of short, sparse and mixed weeds in maize. The red box represents the object detection box. The yellow box represents the area of detail enlargement.
Plants 13 01842 g013
Figure 14. Comparison of segmentation results of model instances under dense maize growth scenario. The red box represents the object detection box. The yellow box represents the area of detail enlargement.
Figure 14. Comparison of segmentation results of model instances under dense maize growth scenario. The red box represents the object detection box. The yellow box represents the area of detail enlargement.
Plants 13 01842 g014
Figure 15. Comparison of segmentation results of model instances in scenarios with normal maize growth and mixed with other plants. The red box represents the object detection box. The yellow box represents the area of detail enlargement.
Figure 15. Comparison of segmentation results of model instances in scenarios with normal maize growth and mixed with other plants. The red box represents the object detection box. The yellow box represents the area of detail enlargement.
Plants 13 01842 g015
Figure 16. Instance segmentation accuracy at different spatial resolutions.
Figure 16. Instance segmentation accuracy at different spatial resolutions.
Plants 13 01842 g016
Table 1. Time cost of manual and SAM-based annotations.
Table 1. Time cost of manual and SAM-based annotations.
MethodAverage Time to Annotate on Particle (s)
Manual annotation28.43
SAM-based semi-automatic annotation5.5
Table 2. Statistics on large, medium and small instances.
Table 2. Statistics on large, medium and small instances.
DatasetClassInstancesTarget Amount
SmallMediumLarge
Train SetMaize19,310162093116,759
Val SetMaize47543891974168
Test SetMaize29,2262278136025,588
Table 3. Model training parameters.
Table 3. Model training parameters.
Parameter NameParameter ValueParameter NameParameter Value
Epochs200Batch Size2
Momentum0.937Weight_Decay0.0005
Table 4. Comparison of segmentation accuracies of different band combination data and different models.
Table 4. Comparison of segmentation accuracies of different band combination data and different models.
ModelBandBoxSegParameters
(MB)
mAPval50mAPval50-95mAPval50mAPval50-95
YOLOv8mNRG0.9520.7940.940.61827.240
NER0.9510.7930.940.615
RGB0.9490.7690.9320.567
YOLOv5mNRG0.9520.7880.9420.61126.531
NER0.9520.7860.9410.609
RGB0.9470.7550.9330.56
PointRendNRG0.9180.6650.8970.52455.755
NER0.9180.6660.9060.524
RGB0.9090.6310.8950.463
Mask Scoring
R-CNN
NRG0.9060.6390.8910.49960.230
NER0.9130.6380.8920.498
RGB0.9040.60.880.435
Mask R-CNNNRG0.9100.6370.8950.51143.971
NER0.9150.6400.9020.515
RGB0.9030.5960.8840.454
Cascade Mask R-CNNNRG0.8050.6190.7820.43777.021
NER0.7770.6050.7640.429
RGB0.7660.5760.3790.752
Table 5. YOLOv8 model (n, s, m, l, x) in different bands’ instance segmentation results.
Table 5. YOLOv8 model (n, s, m, l, x) in different bands’ instance segmentation results.
ModelBandBoxSegSpeed
(ms)
Parameters
(MB)
FLOPs
(G)
mAPval50mAPval50–95mAPval50mAPval50–95
YOLOv8nNRG0.9420.750.9280.57611.73.26312.1
NER0.9410.7510.9280.57912.4
RGB0.9370.7180.9190.52414.1
YOLOv8sNRG0.9470.7780.9370.60420.911.79042.7
NER0.9490.7770.9350.60123.4
RGB0.9460.7440.9310.55122.3
YOLOv8mNRG0.9520.7940.940.61830.827.240110.4
NER0.9510.7930.940.61530.0
RGB0.9490.7690.9320.56731.1
YOLOv8lNRG0.9540.8070.9410.62336.245.937220.8
NER0.9530.8060.9430.62236.1
RGB0.950.7780.9360.57538.6
YOLOv8xNRG0.9520.8110.9420.62459.971.752344.5
NER0.9530.8120.9430.62659.6
RGB0.9520.790.940.58458.8
Table 6. Comparison of accuracy of different instance segmentation models on test set (NRG images).
Table 6. Comparison of accuracy of different instance segmentation models on test set (NRG images).
ModelBoxSegSpeed
(ms)
FLOPs
(G)
Parameters
(MB)
mAPtest50mAPtest50-95mAPtest50mAPtest50-95
YOLOv8m0.7950.5850.7880.46225.2110.427.240
YOLOv5m0.7720.5540.7630.42826.195.426.531
PointRend0.730.4280.7140.34855.590.29955.755
Mask Scoring
R-CNN
0.7250.4110.6980.32747.6183.29660.230
Mask R-CNN0.7010.3890.6820.32266.7145.40843.971
Cascade Mask R-CNN0.6020.3870.5820.28543.5240.6477.021
Table 7. Example segmentation accuracy at different spatial resolutions.
Table 7. Example segmentation accuracy at different spatial resolutions.
ModelBandGSD (cm/Pixel)BoxSeg
mAPval50mAPval50–95mAPval50mAPval50–95
YOLOv8mNRG0.4000.9520.7940.940.618
0.4440.9520.7940.9450.62
0.5000.950.7840.9380.61
0.5330.9480.7830.9390.605
0.5710.9490.7780.9380.604
0.6150.9440.7730.9350.598
0.8000.940.750.9310.573
1.0000.9330.7220.9210.548
1.1430.9250.6960.9130.526
1.3330.9140.6680.8970.494
1.6000.9010.6230.8640.442
2.0000.8760.5620.7930.356
2.6670.8230.4640.6160.232
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Geng, T.; Yu, H.; Yuan, X.; Ma, R.; Li, P. Research on Segmentation Method of Maize Seedling Plant Instances Based on UAV Multispectral Remote Sensing Images. Plants 2024, 13, 1842. https://doi.org/10.3390/plants13131842

AMA Style

Geng T, Yu H, Yuan X, Ma R, Li P. Research on Segmentation Method of Maize Seedling Plant Instances Based on UAV Multispectral Remote Sensing Images. Plants. 2024; 13(13):1842. https://doi.org/10.3390/plants13131842

Chicago/Turabian Style

Geng, Tingting, Haiyang Yu, Xinru Yuan, Ruopu Ma, and Pengao Li. 2024. "Research on Segmentation Method of Maize Seedling Plant Instances Based on UAV Multispectral Remote Sensing Images" Plants 13, no. 13: 1842. https://doi.org/10.3390/plants13131842

APA Style

Geng, T., Yu, H., Yuan, X., Ma, R., & Li, P. (2024). Research on Segmentation Method of Maize Seedling Plant Instances Based on UAV Multispectral Remote Sensing Images. Plants, 13(13), 1842. https://doi.org/10.3390/plants13131842

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop