Next Article in Journal
Estimating Landslide Surface Displacement by Combining Low-Cost UAV Setup, Topographic Visualization and Computer Vision Techniques
Next Article in Special Issue
Sensitivity of LiDAR Parameters to Aboveground Biomass in Winter Spelt
Previous Article in Journal
Fast Marching Techniques for Teaming UAV’s Applications in Complex Terrain
Previous Article in Special Issue
Comparison between Field Measured and UAV-Derived Pistachio Tree Crown Characteristics throughout a Growing Season
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Automated Rice Phenology Stage Mapping Using UAV Images and Deep Learning

1
College of Biosystems Engineering and Food Science, Zhejiang University, Hangzhou 310058, China
2
Huanan Industrial Technology Research Institute, Zhejiang University, Guangzhou 510700, China
3
College of Mechanical and Electrical Engineering, Xinjiang Agricultural University, Urumqi 830052, China
4
Institute of Agricultural Economics and information, Jiangxi Academy of Agricultural Sciences, Nanchang 330200, China
*
Author to whom correspondence should be addressed.
Drones 2023, 7(2), 83; https://doi.org/10.3390/drones7020083
Submission received: 29 December 2022 / Revised: 17 January 2023 / Accepted: 23 January 2023 / Published: 25 January 2023
(This article belongs to the Special Issue UAS in Smart Agriculture)

Abstract

:
Accurate monitoring of rice phenology is critical for crop management, cultivars breeding, and yield estimating. Previously, research for phenology detection relied on time-series data and orthomosaic and manually plotted regions, which are difficult to automate. This study presented a novel approach for extracting and mapping phenological traits directly from the unmanned aerial vehicle (UAV) photograph sequence. First, a multi-stage rice field segmentation dataset containing four growth stages and 2600 images, namely PaddySeg, was built. Moreover, an efficient Ghost Bilateral Network (GBiNet) was proposed to generate trait masks. To locate the trait of each pixel, we introduced direct geo-locating (DGL) and incremental sparse sampling (ISS) techniques to eliminate redundant computation. According to the results on PaddySeg, the proposed GBiNet with 91.50% mean-Intersection-over-Union (mIoU) and 41 frames-per-second (FPS) speed outperformed the baseline model (90.95%, 36 FPS), while the fastest GBiNet_t reached 62 FPS which was 1.7 times faster than the baseline model, BiSeNetV2. Additionally, the measured average DGL deviation was less than 1% of the relative height. Finally, the mapping of rice phenology was achieved by interpolation on trait value–location pairs. The proposed approach demonstrated great potential for automatic rice phenology stage surveying and mapping.

1. Introduction

Rice cultivation has a long history, and it is one of the world’s main cereal crops, as well as a primary source of food for more than three billion people worldwide [1]. The requirements of rice vary according to its growth stage, and different management techniques are required at various stages, such as drying the fields at the late tillering stage to elongate plant stems [2]. Moreover, for rice breeding, the importance and reliability of key traits vary from stage to stage. For example, the rice leaf area index (LAI) should be estimated after the jointing stage [3], while yield should be estimated at the flowering stage or later [4,5]. The phenology in different seasons and the duration of each stage will influence the growth of rice as well [6,7]. Therefore, to cultivate rice well, monitoring the rice growing stage is important, especially for field planting.
Field inspections by workers, even with the aid of a detection model using handheld camera images, are not practical for large areas [8]. A growth simulation model based on environmental parameters such as thermal time accumulation is not stable and accurate [9]. With the trend towards intensive farmland management, remote sensing using satellites or drones offers a highly efficient method of investigation. Satellites can observe a larger area at the scale of a county or province [10]. Moreover, a synthetic aperture radar (SAR) is used for estimating the historical rice growing and planting pattern [11]. Most of these methods rely on time-series data to identify temporal responses of the entire season, which are hard to realize in real-time surveying [12]. Additionally, the temporal–spatial resolution of satellites (greater than 10 m and five days) is too coarse for use in precision agriculture [13].
Drones are widely used in plant phenotype, including crop identification [14], crop biomass estimation [15,16], nutrition assessment [17], seedling counts [18], and also phenology investigation. Compared with infrared imaging capturing the thermal irradiance of the target [19], high-resolution RGB and spectral imaging are more effective for observing detailed rice phenological traits. Yang et al. [2] constructed digital surface models (DSMs) based on UAV images to estimate rice plant height and the growth stage, which could help guide the automated irrigation system for water-saving. Various features, including the vegetation index, color space, and textures, were extracted from regions of interest (ROI) in orthomosaic maps [20]. These features were subsequently processed using ensemble machine-learning algorithms for phenology detection. Ma et al. [21] utilized the red-edge chlorophyll index (CIred edge) and normalized difference vegetation index (NDVI) to monitor the hybrid rice’s initial heading stage. The results verified that the CIred edge is more suitable for monitoring the phenological stage. Additionally, there have been numerous studies that have utilized UAV images to quantify important rice growth characteristics, such as the plant area index (PAI) [22], leaf dry biomass (LDB) [23], and storage organ biomass [24]. However, many of these studies have relied on time-series (multi-temporal) vegetation index (VI) data for the rice growth stage, which are not robust for expansion or generalization. Yang et al. [25] proposed a convolutional neural network (CNN) that utilizes mono-temporal UAV images for detecting rice phenological stages and received an accuracy of 83.9% with an auxiliary regional mean thermal time input. Despite its promise, this classification model relies on manually plotted ROIs and requires the generation of orthomosaic maps before phenology detection.
Deep learning has demonstrated superior performance in pattern recognition [26,27,28], and the multi-class image segmentation model has the potential for fully automatic target extraction and mapping [29]. Researchers Lan et al. [8], Deng et al. [30], and Sai et al. [31] have developed several semantic segmentation networks for real-time identification and mapping of weeds in paddy fields using UAV. In [32], a paddy field segmentation model was proposed that combines an attention mechanism with an adaptive spatial feature fusion algorithm based on DeepLabv3+. This proposed model, referred to as SA-DeepLabv3+, attained a higher accuracy and speed. To identify the rice lodging area, Yang et al. [33] used an FCN-AlexNet network and added an extra vegetation index map to the input image to identify areas of rice lodging. However, there have been few studies on rice field area segmentation and multi-stage discrimination.
To address these challenges, the objectives of this study are (1) to create a dataset for rice field segmentation and phenology classification; (2) to develop a fast and accurate model for multi-class segmentation and compare it to existing models; (3) to investigate an effective and reliable workflow for locating and mapping traits using UAV images sequence; (4) to test the effectiveness of the proposed system with a rice phenology mapping test.

2. Materials and Methods

2.1. Experimental Sites and Equipment

To ensure data diversity, unmanned aerial vehicle (UAV) images of rice fields were collected from 14 flight campaigns in the main producing provinces of Zhejiang, China, from June 2021 to September 2022. These campaigns were conducted at 9 different experimental fields, including djd1~djd5, qt1, qt2, sds, xs2 in Hangzhou and lq1, lq2, ra1, ra2 in Wenzhou City, as shown in Figure 1. On the same day of the UAV flight, rice planting experts investigated and recorded the ground truth stage in the field.
All aerial images were collected by a Zenmuse P1 (DJI Technology Co., Ltd., Shenzhen, China) full-frame camera mounted on the Matrice 300 RTK quadcopter (DJI Technology Co., Ltd., Shenzhen, China) through a three-axis gimbal at the vertical overhead view. The camera was equipped with a 35 mm focal length lens, and its sensor size, photo size, and pixel size are 35.9 × 24   m m , 8192 × 5460 , and 4.4   μ m , respectively [34]. Other equipment includes an i70II handheld RTK (CHCNAV, Shanghai, China) for ground control point (GCP) locating, a server with two Nvidia RTK-2080Ti GPUs for deep learning model training, and a computer with Nvidia RTX-2070 GPU for model speed test and other data processing.
As for the software environment, Pix4D Mapper 4.4.12 (Pix4D SA, Prilly, Switzerland) software was used to generate orthomosaics through SFM and calibrate camera parameters [35]. QGIS 3.24 (OSGeo, Chicago, IL, USA) software was used to visualize the waypoints and orthophotos for each flight and also for interpolating and mapping the distribution of the trait [36]. To annotate segmentation mask labels effectively, we adopted the interactive annotation software EISeg (Baidu, Beijing, China) [37]. Additionally, the construction of deep learning models was based on the PyTorch framework and MMSegmentation toolbox [38], and other data processing and algorithm were implemented with Python 3.8 scripts.

2.2. Construction Process of Dataset

To provide a reliable data basis for rice filed segmentation and phenology classification of UAV images, we used this section to illustrate the flow and techniques of building the Paddy Segmentation dataset (PaddySeg), as shown in Figure 2.
All UAV flights were at 25   m relative height above ground with the camera lens vertical to the ground, such that the ground sampling distance (GSD) of the original P1 image was around 0.31   c m / p i x e l . During each flight, the onboard real-time kinematic system (RTK) was used to acquire high positioning accuracy of horizontal 1   c m [39]. The forward and side overlap ratios were both 70% to ensure abundant feature matching at the structure-from-motion (SFM) process, and the average flight speed was 2.5 m/s. There are no strict limitations on weather or illumination, and the camera was set to P mode of automatic exposure to facilitate the diversity of the image data and robustness in subsequent model training. The waypoints and routes were planned automatically after selecting the target field. Moreover, all images were retrieved back from the SD card onboard after the flight campaign.
In most cases, the original size of the aerial image ( 8192 × 5460 ) is too large to process in deep learning models, causing out-of-memory in training and high latency during testing. Moreover, not all pixels in the image are worth calculating, considering the high resemblance in the farmland scene and the high overlap of aerial images. Moreover, smaller patches of the image are more convenient and fluent at labeling.
Consequently, each image was first divided into 5 rows and 5 columns of patches of 1638 × 1092 pixels (around 5.1 m × 3.4 m field of view), and patches were selected manually to eliminate similar (Sim-Elim) or redundant features. For example, patches containing more than one period of rice and other plants or weeds are preferred, while patches full of the crop are undesirable. After this process, the number of patches was reduced from N * r * c to M , where N is the original image number of one mission and M is the target patches number according to the data diversity of the mission.
According to the management demands and key feature differences, we divided the rice phenology into 4 stages: 1. seedling: from seedling to the tillering stage, around 25 days after transplanting; 2. jointing: from jointing to booting stage, around 30 days; 3. heading: from heading to the flowering stage; and 4. filling: from filling to ripening stage. Each stage has its unique phenotype. At the seedling, bare land can be observed between the rice plants of one row, while this window is closed at the jointing stage when the grown-ups cover the land. Thanks to the high resolution of the P1 camera, we could observe the rice flowers in the photo of the heading stage and also the rice ear in the filling stage.
As the right part of Figure 2 shows, apart from the rice region, other vegetation, bare land, field ridges, or weeds were considered as “others”, which is labeled as 0. We used EISeg to label the foreground paddy region and phenology stage by adding positive and negative sample points and generating corresponding masks. Table 1 shows the sample number, class pixel number, and distribution across the different missions. Although some of the campaigns have only one plot of rice (and only one stage in the mission), others contain multi-stages, especially for djd and qt sites where rice seedlings were transplanted at different seasons. It is worth noting that the total pixel numbers of each class are imbalanced, which may lead to majority bias during training [40], and we will discuss this later.
After gathering all 2600 pairs of image-mask, we split the dataset into 3 parts randomly with a constant seed 2022. The sample number and ratio of training, testing, and validation are 1820, 520, 260, and 7 : 2 : 1 .

2.3. Multi-stage Rice Field Segmentation Model

Several studies and experiments demonstrated the efficiency of bilateral structure in semantic segmentation tasks [8,41]. With a detailed branch collecting low-level spatial details and a semantic branch extracting high-level semantics, this kind of model greatly accelerates the inference while improving the segmentation performance [41]. Due to the amount of aerial data and real-time inferencing demands, a faster model is required. However, the original detailed branch and decoder head of BiSeNetV2 demand heavy computational expenses of nearly 90%, as shown in Table 2 [42].
Consequently, we implement an efficient ghost convolution module and construct the Ghost Bilateral Network (GBiNet) by improving the computational bottlenecks in BiSeNetV2. In Section 2.3.1, the ghost convolution module and other basic components are introduced. Section 2.3.2 describes the overall structure of GBiNet, and other experimental settings and parameters are detailed in Section 2.3.3.

2.3.1. Ghost Convolution Module (GCM)

There are many similar and redundant feature maps in a well-trained convolutional neural network to ensure a comprehensive and stable understanding of the image data. As shown in Figure 3, after the first stage of the Detail Branch in BiSeNetV2, the output feature maps contain many similar pairs or groups. In each group, one feature map may convert to another through a cost-effective linear transformation [43].
Figure 4a illustrates the process of ghost convolution, and it can be denoted as:
Y 0 = C o n v 3 × 3 X ,   s , c ,
Y = C o n c a t Y 0 ;   G L i n e a r Y 0 ,   r
where X c × h × w are the input feature map of c channels and h × w size, C o n v 3 × 3 is the vanilla convolution with 3 × 3 kernel size, s stride, and c kernels, Y 0 c × h s × w s is the initial feature map of c channels and h s × w s size, and G l i n e a r is the group linear transformation that generates r 1 ghosts feature maps from each channel of Y 0 , and r N * is the ratio of ghost feature maps at output Y .
As output channels are normally predefined, the initial channels are dependent on r , and w = w o r . It is worth noting that the computation cost of G l i n e a r is much smaller than conventional convolution, and the ratio r indicates a speed-up ratio compared to ordinary convolution, as proven in [44]. Although a large r ratio may bring computation costs down, it can lead to instability for lacking cross-channel information fusion.
Taking advantage of GCM, Ghost Block (G-Block) is introduced for the Detail Branch of the bilateral segmentation network. Each G-Block starts with a GCM of stride 2 to reduce computation cost, and is followed by one or two GCM of stride 1 to deepen the feature extraction network, as shown in Figure 4b. the structure is simple yet effective for Detail Branch with a wide feature dimension. Moreover, we also noticed another computation-heavy part of BiSeNetV2 is the FCN head [45], so a Ghost Convolution Network segmentation Head (GCN-Head) is designed, as shown in Figure 4c. The GCM of 1 stride generates abundant feature maps first, then depthwise convolution C o n v p t compresses the channel number to the number of classes. After a dropout layer to avoid overfitting, feature maps are upsampled to the input shape by bilinear interpolation.

2.3.2. Ghost Bilateral Network (GBiNet)

The proposed Ghost Bilateral Network (GBiNet) is based on the BiSeNetV2 [42], and the overview of GBiNet is shown in Figure 5. There are four main parts: Detail Branch, Semantic Branch, Aggregation Layer, and Segmentation Head for inference. Moreover, a set of Auxiliary Segmentation Heads works in the training phase only.
A bilateral network has two different encoder branches. The Detail Branch requires high channel capacity to encode rich spatial details. Meanwhile, as this branch focuses only on low-level features, a shallow and wide-layer structure is preferred. On the contrary, Semantic Branch aims to capture high-level semantics, such that a deep and narrow layer structure is needed. Because of the different shapes of output feature maps from two parallel branches, the bilateral guided aggregation (BGA) block is used to merge the two complementary features. The Auxiliary Segmentation Heads are inserted at different stages of the network, which can enhance the feature representation during the training phase while adding no computation costs in the inference stage.
According to the floating-point operations (FLOPs) of each component, the Semantic Branch and BGA Block consume less than 10% of the overall computation cost in the original BiSeNetV2 model. The fast down-sampling strategy, lightweight depthwise convolution, and global pooling embedding have been adopted in the original stem block (Stem-Block), gather-and-expansion layer (GE-Block), and context embedding block (CE-Block), respectively, in the original Semantic Branch. Additionally, the GBA block uses the context information of Semantic Branches to guide the feature response of Detail Branches. Different scale feature representations can be captured through different scale guidance, which enables effective communication between two branches. So, the components and structures of these two parts are kept in our GBiNet. For details, please refer to [42].
The other two parts, Detail Branch and Segmentation Head, are the bottlenecks of speeding up model inferencing. Therefore, we have improved both parts with GCM to construct a more efficient model. Table 3 depicts the overall architecture of standard GBiNet. The Detail Branch mainly consists of a stack of G-Block with GCM as the basic unit. Each block implies a stage according to the sizes of input feature maps. All GCM are applied with s t r i d e = 1 , except that the first one in each stage or G-Block is with s t r i d e = 2 . The implicit kernel channels of each stage are 64 ,   64 ,   128 , and the increase of the channels is also at the first GCM. The last output channel number of the Detail Branch is the same as that of the Semantic Branch, whose implicit kernel channels are 16 ,   32 ,   64 ,   128 , which indicates the bandwidth from the encoder to the decoder. The divergent strides combination of the two branches benefits multi-scale feature extraction for semantic segmentation tasks. After bilateral aggregation, a GCM extends the features to 1024 decode channels, and a pointwise convolution layer with dropout is utilized to decode feature maps to a 5-channel mask for segmentation. An interpolation layer of 8-times upsampling is also applied to yield the final mask, as pictured in Figure 4c.
We have also designed a tiny model, namely GBiNet_t. The output channels of its Detail and Semantic Branch are 16 ,   16 ,   32 and 8 ,   16 ,   16 ,   32 . Additionally, all G-Blocks at the Detail Branch are of strides 2 ,   1 , and the default number of decode channels is 32 for GBiNet_t.

2.3.3. Experimental Setup and Parameters

To evaluate the performance of the proposed models on the multi-stage rice field segmentation tasks, we have trained GBiNet and other existing models on the PaddySeg dataset. All models are trained from scratch with the Kaiming initialization method [46]. For all training schedules, the optimizer is stochastic gradient descent (SGD) with a 0.9 momentum and 5 e 4 weight decay. The initial learning rate is 5 e 3 and is decayed using a poly strategy with 0.9 power and 1 e 4 minimum rate. The max training iterations is 60 k , which is abundant for all models to fully fitted on PaddySeg, and the training weights would be saved and evaluated every 6 k iterations. The best-trained weights on the evaluation dataset were kept for testing.
For evaluation, mean intersection over union (mIoU) is the primary metric. Meanwhile, overall pixel accuracy (aAcc) and intersection over union (IoU) of each class were also recorded for balance assessment over multi-classes. The model weight’s number and FLOPs were measured to evaluate the memory and computation costs.
For training data augmentation, we randomly resize, randomly crop, and randomly horizontally flip the images to a target size. The scale range of resizing is 0.5 ~ 2.0 , and 0 and 255 would pad the absent area of the image and its mask. For the test pipeline, no augmentation method is adopted. There are two kinds of data processing modes, sliding window (Slide) and down-sampling whole (DSW). The first mode has a target size of 546 × 546 , and during inference, a 546 × 546 sampling window would Slide on the 1638 × 1092 image with 364 ,   364 stride and yield 6 patches to feed the model. While DSW has an 819 × 546 target size, it down-samples the image to the target size and feeds the whole image to the model. The DSW input mode effectively simulates images taken from twice the height (50 m) compared with the Slide mode (25 m) and larger ground sampling distance (GSD). All random processes used the random seed of 2022 to ensure reproducibility. The model implementation code was arranged in MMSegmentation style under PaddySeg/GBiNet/directory at (https://github.com/HobbitArmy/PaddySeg, accessed on 3 January 2023).
All training progress was conducted on the server with two RTX-2080Ti GPUs, and each GPU takes 4 samples per iteration; thus, the overall batch size is 8 . To test the practical inference speed, we measured frames per second (FPS) on the PC with one RTX-2070 GPU. We fed the test set (520 samples) to the pending model and recorded its processing FPS, then repeated this 5 times to eliminate fluctuation.

2.4. Traits Locating and Mapping System

To achieve the fast locating and mapping of crop traits, we used this section to provide a system to process the original UAV images directly, which does not require structure-from-motion (SFM) and image mosaic. The distribution map of target traits can be obtained, combined with the feature extraction methods demonstrated in the previous section. Section 2.4.1 introduces the geo-locating method of every pixie in a UAV image based on a simplified photogrammetry model. In Section 2.4.2, we split the UAV image into patches in sequence and discard some patches to achieve the sparse geo-distribution.

2.4.1. Direct Geo-Locating (DGL)

According to photogrammetry, to calculate the absolute location of the target point in an image, the coordinate transformation relationship between image points in different spatial rectangular coordinate systems should be established first [47]. This can be complicated for UAV photographs, where a lot of exterior elements are required to achieve multiple transformations. Photos are transformed from image plane to image space, then to the gimbal and the aircraft body, and finally, to a geodetic coordinate system (GCS).
Fortunately, with a three-axis gimbal, vertical photographs were taken with the optical axis of the camera vertical to the ground [48]. As Figure 6a illustrates, the roll or pitch of the aircraft would not cause deflection of the camera, and only rotational movement leads to the yaw angle changing of the camera [49]. Additionally, the distortion of the lens was calibrated, and the RTK position recorded in the image EXIF information is accurate and was calibrated from the RTK antenna to the principal point of the camera. Based on this, for a flat rice field, we simplify the imaging process to a central projection model with the subject plane G - X Y parallel to the camera sensor plane o - x y , as shown in Figure 6b. Using the internal and external parameters of the camera and the collinear equation, the absolute location of the target point in a photo can be direct geo-located (DGL).
More specifically, for GCS G - X Y Z , y-axis points to the north, x-axis points to the east, and the z-axis points upwards. For the auxiliary imaging coordinate system o - x y z , o is the central point of imaging film, the y-axis points to the top of the image, the x-axis to the right, S is the exposure station (also lens center) at X s ,   Y s ,   Z s , N is the vertical ground projection point, S - N is the optical axis position, the length of S - o is the focal length, and κ is the counterclockwise angle from the y-axis to the north (y-axis). The target point T is collinear with the corresponding imaging point t and the projection center S , and α denotes the angle from the y-axis to the vector o t . The coordinate of T at G - X Y plain can be calculated from the following equation:
c o o r d T = c o o r d N + N T = X s , Y s + N T
N T = S N S o * o t = S N S o * o t * sin κ α , c o s κ α
where S N is the relative height of the exposure station, S o is the focal length, o t is the distance between the central point and the target point on the camera sensor, which can be calculated by:
o t = p i x e l d i s * s i z e s e n s o r p i x e l s e n s o r
where p i x e l d i s is the pixel distance between the central point and target point, s i z e s e n s o r is the effective sensor size of the camera (along the width or length), and p x i e l s e n s o r is the corresponding pixel number.
In practical use, we have adopted the calibrated s i z e s e n s o r and focal length yielded by the Pix4D project rather than the labeled value. For our P1 camera, the average focal length at 25   m relative height is 34.90   m m , and the sensor size is 35.0000 × 23.3276   m m . These three values are fixed after initialization, and other parameters are loaded from metadata in the image. For convenience, we implemented the DGL method with Python and packed it up into class. See the code DirectGL.py for more details.
The ground control points (GCP) were set and located using a high-accuracy handheld RTK to evaluate the locating accuracy of DGL, as shown in Figure 7a,b. Then we used a UAV to scan the area and load the UAV image. In Figure 7c,d, given the measured G C P l o c , the predicted GCP location G C P p r e d on the image is plotted with a blue dot. By selecting the observed G C P o b s manually, the haversine distance between G C P p r e d and G C P o b s is calculated and recorded.

2.4.2. Incremental Sparse Sampling (ISS)

Due to the high overlap of UAV images, repeated calculation of the same region is unavailing and brings high latency to the system. This section supplies an incremental sparse sampling method for UAV image sequence, which splits images one by one into patches and discards the redundant ones.
As depicted in the left side of Figure 8, the whole UAV image is split into 5 rows and 5 column patches first. Secondly, geo-locations of the central 3 × 3 candidate patches are calculated (considering the DGL-acc, the outer edge patches of each image are removed). Patches of the first image in sequence would be accepted directly and stored. The distance matrix between the subsequent patches set of one image and the stored patches is calculated. Any candidate patch that is too close to the stored patches ( < d i s p a t c h * r e l m ) would be excluded, where d i s p a t c h is the short distance between two patches in one image, and r e l m is the distance ratio.
Because the stored patches are dynamically changing, and patches to be stored are selected step by step, we call this method incremental sparse sampling (ISS). ISS method is summarized in Algorithm A1 of Appendix A and implemented in the PatchSparseSampling.py code file.
As shown on the right side of Figure 8, after sparse sampling, the label mask is computed for each stored patch using a segmentation model. Moreover, the mask is averaged pooled to down-sampling the trait’s value, which is like splitting the patch into smaller boxes, and the pooling kernel size is the box size. This operation reduces data redundancy while improving robustness. The image location of every box is recalled to DGL their geo-location (see code Patch2Distribution.py for detail). With the traits value-location pairs, the rice phenology distribution map is acquired using inverse distance weighted (IDW) interpolation.

3. Results

In this part, we first report the performances of GBiNet and current classic semantic segmentation models on the PaddySeg dataset in Section 3.1. Then, in Section 3.2, the accuracy of direct geo-locating is tested. Finally, Section 3.3 shows a practical test of rice phenology distribution mapping that integrates all methods for practical use.

3.1. Segmentation Model Performance

According to the models, training methods, and hyper-parameters described in Section 2.3.3, a series of training experiments were conducted. Four classical semantic segmentation networks were trained and evaluated on PaddySeg, as summarized in Table 4. Apart from the segmentation accuracy indicators mean intersection over union (mIoU) on validation and test set, the costs of processing on memory, computation, and time are reflected by weights number (in Million units), floating-point operations (FLOPs, in Giga unit) and frame per second (FPS), respectively. Towards real-time aerial image processing, speed and accuracy are the key parameters. Apart from the sliding-window (Slide) input mode, an extra experiment of the down-sampling whole (DSW) input mode was conducted.
At Slide input mode, all models achieved a mIoU on the test set above 91%, with the highest pspnet reaching 91.57%. At the same time, the inference speed of BiSeNetV2 (5.4 FPS) exceeds other models with an average of 2.7 FPS. However, this level of latency is still too high for practical use. Fortunately, all models acquired a six- to nine-times boosting speed from DSW input mode, as shown in the last column of Table 4. Moreover, the price is a minor performance reduction according to the 0.36% test set mIoU drop of BiSeNetV2 with DSW mode, as revealed in Table 5. Therefore, in the following experiment, we only adopted the DSW input mode.
The red and cyan lines in Figure 9a show the convergence process of the BiSeNetV2 model with Slide and DSW input mode. DSW mode leads to a much lower loss value for training, while their average accuracies on the validation set are almost the same (Figure 9b). This implies that coarse images are more easily segmented by the model, resulting in lower training loss values.
To explore the potential of the proposed method comprehensively, we designed multiple GBiNet and its tiny sibling GBiNet_t. As Table 5 depicts, the number after _ r suffix denotes the ghost ratio, the number right after G B i N e t _ or G B i N e t _ t is the last output channels (LOC) of the Detail and Semantic Branches for tuning the width of model, and the number after d x is the expanding ratio of decode channels in GCN-Head. Accordingly, if the LOC number is 64, then the output channels of its Detail and Semantic Branches are {32, 32, 64} and {16, 32, 32, 64}; if the number is 32, then the output channels are {16, 16, 32} and {8, 16, 16, 32}. If not specified, all parameters are set as detailed in Section 2.3.2 by default. For example, GBiNet_r2 has {64, 64, 128} and {16, 32, 64, 128} output channels of the two branches, 1024 decode channels, and the ghost ratio of two, while GBiNet_t32dx2_r4 has {16, 16, 32} and {8, 16, 16, 32} output channels, 64 decode channels, and ghost ratio of four.
From the results in Table 5, the speed of the proposed models surpassed all existing models. When the model ghost ratio increased from two to eight, the FLOPs reduced significantly, and the speed rose accordingly, while the accuracy decreased as expected. In particular, the GBiNet_r2 performed even slightly better than the original BiSeNetV2, achieving the highest 91.5% mIoU-test performance and lower FLOPs. According to Figure 9, the convergence curves of GBiNet_r2 and BiSeNetV2 almost coincide, verifying the robustness of the ghost convolution. Furthermore, as we changed the LOC and decode channel number smaller, the speed of the model increased significantly, boosting the tiny model GBiNet_t32dx2_r4 to a speed approaching 62 FPS with an acceptable mIoU around 90.2% on the test set.
The ground truth and segmentation results of the most accurate and fast GBiNet model are shown in Figure 10. Some spots in the ground truth mask of the last column are weeds among rice plants which are hard to be identified by models.

3.2. Direct Geo-Locating Accuracy

Although direct geo-locating (DGL) is easy to achieve in principle, this section gives the result of actual locating experiments to verify its feasibility. Many aspects of reasons can lead to the deviation of final positioning, including inaccuracy of the airborne RTK system, instability of the gimbal, and camera distortion. To assess the DGL system comprehensively, we directly calculated the haversine distance of the true location and predicted location as the deviation.
A total of 20 ground control points (GCP) have been set at three sites with their accurately measured geographic coordinate locations (GCP_loc) using handheld RTK, as shown in Table A1 of Appendix A. Each GCP (with a GCP_name) would be observed by several UAV images (counted as OBS_num), and the distance between GCP_loc measured on the ground and the observed location estimated on the aerial image was recorded. We average all distance deviations of each GCP (DEV_avg) and found that the locating deviation of an arbitrary point is near 0.21 m.
The distributions of all observed deviations are fitted with communally used functions in Figure 11. Moreover, the expectations of these functions are also around 0.2 m, which is less than 1% of the relative flying height. This accuracy level, coupled with smoothing filtering of interpolation during the subsequent mapping process, is sufficient for agricultural field usage.

3.3. Rice Phenology Mapping

After feature extraction and trait locating, an integrated test on rice phenology distribution mapping is presented. Figure 12 depicts the key intermediate diagram generated in the system workflow as described in Section 2.4.2. Each blue star in Figure 12a is a waypoint where an image is taken by the UAV with a red sequence number marked aside. In Figure 12b, these image geo-locations (IGL) are marked as blue stars as well, and the red points are sparsely sampled patch geo-location (PGL). Because the box geo-locations (BGL) in one patch are arrayed one by one, they are displayed like a purple block around the PGL point.
PGL points distribute evenly thanks to the sparse sampling, and the total patch number to be processed further is decreased from 1170 (130 images * nine patches each) to 294. This greatly saved the computation costs. Although most of the target field is covered by purple, some of the regions are ignored mistakenly, as shown blank in Figure 12b.
Based on the BGL locations and corresponding trait values (rice phenology of the box area), we used inverse distance weighted (IDW) interpolation in QGIS software and generated a heat map of rice phenology distribution, as Figure 12c shows. The color mask represents others, seedling, jointing, heading, and filling from blue to green to red, with values exchanging from 0 to 4. Areas marked with white numbers from 1 to 4 are rice. We hid the zero value in the figure so that roads, bare soil, trees, and other regions are blank. Moreover, the rice area is also represented by different colors and values according to the phenology stage.
The rice growing period in different plots can be judged with this picture. For example, the field to the north and the roadside strip areas are mainly at tillering stage. While rice planting troughs in the middle are mainly at the filling stage, the east side troughs are mainly at the jointing stage.
In Figure 12c, there is an unusual blue area in the center of the planting troughs area. As we zoom in, as shown in Figure 13a, it is obvious that the low-value region (circled in white) is caused by the presence of weeds. Because of the fine label of PaddySeg, weeds amongst rice fields are also considered to be others, so the model judges this area as non-rice. Another area circled in yellow at the bottom right is the rice area, but it lacks a number mark, which is caused by the incomplete patch coverage mentioned earlier. However, the existence of interpolation makes the final mapping results more robust, where the color of this area still indicates that it is rice at the filling stage (Figure 13b).

4. Discussion

4.1. Efficiency of GBiNet

Bisenetv2 is an efficient real-time semantic segmentation network that employs a bilateral structure in the decoder portion. However, the calculation cost of the detailed branch and decoder portion is too high, accounting for nearly 90%, as shown in Table 6. Therefore, this paper proposed a ghost convolution module to improve and reduce the computation consumption of these two parts. By adopting a ghost ratio of two alone, half of the computation was reduced, and the computation continued to decrease as the ratio increased. In addition, a tiny model called GBiNet_t was also proposed, which compressed the overall computation to less than 0.5 GFLOPs. The tiny model not only reduced the computation but also cut down the number of parameters. Moreover, the smallest model has only 0.091M parameters, which has the potential to provide real-time computation and be deployed on an edge device.
It is not difficult to reduce the network size; the key is how to maintain satisfactory accuracy with simultaneous fast processing. Ghost convolution uses grouped pointwise linear transformations, where group convolution can be seen as a decoupling of the original convolution operation, improving the sparsity between filters in the original convolution operation and playing a regularization role to some extent [50]. This is one of the reasons why GBiNet_r2 performs better than BiSeNetv2.

4.2. Confusion Matrix and Classes Accuracy

As previously mentioned, the number of sample pixels was unbalanced between different categories. The majority class may acquire biased attention from the model during training and reach a higher recall ratio, but this is not the case in our experiments. As shown in Table 7, although the “others” class had the largest number of samples, its overall classification accuracy was the lowest. While the “heading” and “seedling” classes had a smaller number of samples, their accuracy was also at a moderate level.
The minority was not ignored by the model, while the majority class gained no bonus from its number as well. Setting class weights to weaken the majority class of “others” and enhance the minority class of “heading” or “seedling” would exacerbate the imbalance. Therefore, in this study, class weights were not used in the loss function.
The confusion matrix of the GBiNet_r2 model segmentation class accuracy on the test set is shown in Figure 14. Most of the misclassifications are caused by predicting various stages of rice as “others”, with an error rate of around 2% to 4%. One reason is that the weeds amongst rice plants are also labeled as others, which are relatively difficult to identify, even for humans. Another reason is that the areas of rice at various stages are similar. Perhaps when counting the number of categories, all rice areas should have been considered as one category. In this case, others become the minority class, and the model is more biased towards jointing and filling, which is consistent with the results. It is worth exploring whether this kind of class-sample statistics can be used to equalize the model results in future experiments.
Additionally, the misclassification rate between adjacent categories is also relatively high, particularly when the seedling category is misclassified as the jointing category or the heading is misclassified as jointing. This result was expected, as the features of adjacent stages of rice are relatively similar and difficult to be identified.

4.3. Limitation and Future Study

We have proposed a range of methods and approaches for investigating rice phenological stages using UAV images. However, there are still several drawbacks: Firstly, the rice phenological stage in the PaddySeg dataset was not finely divided, only encompassing four key periods. Secondly, to ensure a clear view of key features during labeling, we chose relatively low-flight altitude UAV images, which were unsuitable for monitoring large areas. Thirdly, although the data annotation was semi-automatic, it was still time- and labor-intensive. In addition, the issue of imbalanced inter-class accuracy required further investigation and experimentation.
For future research, the direct processing workflow of UAV image sequences could be expanded to other application scenarios, including planting and harvesting progress investigations. Meanwhile, data could be collected at a higher altitude for high-efficiency large-area surveys. In an aspect of the application, we plan to seek more efficient and resolution-acceptable flight settings to facilitate the large-scale monitoring of crop traits [51], explore the feasibility of using super-resolution methods [52], enhance UAV images and downstream models, and unify UAV images at multi-levels of ground sampling ratio. For locating and mapping, we will focus on the local SFM or satellite map matching [53] methods to assist in the photogrammetric positioning of target points in a single UAV image. In terms of modeling, we will attempt to introduce super-pixel methods [54] for partitioning images first and combine this prior to processing with deep learning classification models, which will further improve the speed rather than segmentation models.

5. Conclusions

In this paper, to achieve automatic field rice phenology stage investigation, we have built a multi-stage rice field UAV image segmentation dataset, PaddySeg, constructed a Ghost Bilateral Network (GBiNet) for rice field segmentation and phenology classification, and designed an efficient workflow for trait locating and mapping. According to the results on PaddySeg, the most accurate GBiNet_r2 with 91.50% mIoU-test and 41 FPS speed exceeded the baseline model BiSeNetV2 (90.95%, 36 FPS). While the fastest GBiNet_t32dx2_r4 reached over 1.7 times of speed boosting (62 FPS) and kept a similar level of mIoU-test. Additionally, a straightforward and effective workflow was designed, where patches were split from the image incrementally and sampled sparsely (ISS) to eliminate computational redundancy. These patches were fed into the segmentation model that generates traits mask. Based on UAV photogrammetry, each pixel could be direct geo-located (DGL). The locating accuracy was measured on 20 ground control points (GCP) with a 21 cm average deviation (<1% relative height). The final rice phenology mapping was achieved in QGIS with an interpolation of the trait value–location pairs. This phenology mapping system could aid decision-making toward the automatic management of the rice field.

Author Contributions

Conceptualization, F.L. and X.L.; methodology, F.L., X.L. and J.Z.; software, X.L. and R.Y.; validation, F.L., X.L. and J.Z.; formal analysis, X.L., J.Z. and R.Y.; investigation, X.L., Y.L. and R.Y.; resources, F.L., Z.Y. and J.J.; data curation, X.L., Y.L. and J.J.; writing—original draft preparation, X.L., J.Z. and F.L.; writing—review and editing, F.L., X.L. and J.Z.; visualization, X.L., Y.L., J.Z. and R.Y.; supervision, F.L., J.J. and J.Z.; project administration, F.L., Z.Y. and J.Z.; funding acquisition, F.L. and Z.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Science and Technology Department of Guangdong Province (grant number 2019B020216001), the Science and Technology Department of Zhejiang Province (grant number 2020C02016), and the Science and Technology Department of Shenzhen (grant number CJGJZD20210408092401004).

Data Availability Statement

The datasets collected and generated in this study are available upon request to the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Algorithm A1. Incremental Sparse Sampling of UAV Image Patches
Input: { I N × 1 ,     r p ,     c p ,     r e l m ,     e } # I N is N -length image index; r p and c p is the row and column number of patches set; r e l m is the ratio that defines the minimum threshold distance, and e is the edge number of patches to be removed
Output:   G L l  
1: G L l =  
2:for  i  in  e , , r p e  do for j in  e , , c p e  do
3:  add  D G L p ( I 1 , i , j ) to  G L l # D G L p is direct geo-locating of a patch in the image
4:for  k  in  2 , , N  do
5:   G L k =  
6:  for  i  in  e , , r p e  do for j in  e , , c p e  do
7:   add  D G L p ( I k , i , j ) to  G L k
8:   end for
9:   d i s m a t n × m   =   C D M   G L k n , G L l m # C D M is calculating distance-matrix
10:  for  i  in  1 , , n  do for j in 1 , , m do
11:   if  d i s m a t i , j > d i s p a t c h * r e l m  then
12:    append G L k i , j  to  G L l
13:    end if
14:   end for
15:end for
16:return  G L l
Table A1. Direct Geo-Locating Accuracy at 3 Experimental Sites.
Table A1. Direct Geo-Locating Accuracy at 3 Experimental Sites.
IndexGCP_nameGCP_locOBS_numDEV_avg (m)
1lq1_128.21048270,121.0490756070.22
2lq1_228.21044813,121.0488965070.23
3lq1_328.21040046,121.0486898080.24
4lq1_428.21034937,121.0486553040.14
5lq1_528.21036053,121.0485112080.20
6lq2_128.24022942,121.0282729540.21
7lq2_228.24042392,121.0282599140.19
8lq2_328.24053005,121.0282508080.25
9lq2_428.24060837,121.0282462680.20
10lq2_528.24065780,121.0282447880.21
11sds_130.07501281,119.9242652170.32
12sds_230.07498620,119.9242731870.30
13sds_330.07500016,119.92423888130.24
14sds_430.07501029,119.92420711120.23
15sds_530.07498435,119.92420728130.25
16sds_630.07480693,119.92382305120.26
17sds_730.07480502,119.9237867180.16
18sds_830.07485181,119.9238029880.16
19sds_930.07489651,119.92381531120.18
20sds_1030.07489462,119.9237842780.13
Sum: 166Avg: 0.21

References

  1. Yu, S.; Ali, J.; Zhou, S.; Ren, G.; Xie, H.; Xu, J.; Yu, X.; Zhou, F.; Peng, S.; Ma, L.; et al. From Green Super Rice to Green Agriculture: Reaping the Promise of Functional Genomics Research. Mol. Plant 2022, 15, 9–26. [Google Scholar] [CrossRef]
  2. Yang, C.-Y.; Yang, M.-D.; Tseng, W.-C.; Hsu, Y.-C.; Li, G.-S.; Lai, M.-H.; Wu, D.-H.; Lu, H.-Y. Assessment of Rice Developmental Stage Using Time Series UAV Imagery for Variable Irrigation Management. Sensors 2020, 20, 5354. [Google Scholar] [CrossRef]
  3. Wang, L.; Chen, S.; Peng, Z.; Huang, J.; Wang, C.; Jiang, H.; Zheng, Q.; Li, D. Phenology Effects on Physically Based Estimation of Paddy Rice Canopy Traits from UAV Hyperspectral Imagery. Remote Sens. 2021, 13, 1792. [Google Scholar] [CrossRef]
  4. Wang, F.; Yao, X.; Xie, L.; Zheng, J.; Xu, T. Rice Yield Estimation Based on Vegetation Index and Florescence Spectral Information from UAV Hyperspectral Remote Sensing. Remote Sens. 2021, 13, 3390. [Google Scholar] [CrossRef]
  5. Zhou, J.; Lu, X.; Yang, R.; Chen, H.; Wang, Y.; Zhang, Y.; Huang, J.; Liu, F. Developing Novel Rice Yield Index Using UAV Remote Sensing Imagery Fusion Technology. Drones 2022, 6, 151. [Google Scholar] [CrossRef]
  6. Guo, Y.; Fu, Y.; Hao, F.; Zhang, X.; Wu, W.; Jin, X.; Robin Bryant, C.; Senthilnath, J. Integrated Phenology and Climate in Rice Yields Prediction Using Machine Learning Methods. Ecol. Indic. 2021, 120, 106935. [Google Scholar] [CrossRef]
  7. Ge, H.; Ma, F.; Li, Z.; Du, C. Grain Yield Estimation in Rice Breeding Using Phenological Data and Vegetation Indices Derived from UAV Images. Agronomy 2021, 11, 2439. [Google Scholar] [CrossRef]
  8. Lan, Y.; Huang, K.; Yang, C.; Lei, L.; Ye, J.; Zhang, J.; Zeng, W.; Zhang, Y.; Deng, J. Real-Time Identification of Rice Weeds by UAV Low-Altitude Remote Sensing Based on Improved Semantic Segmentation Model. Remote Sens. 2021, 13, 4370. [Google Scholar] [CrossRef]
  9. Gao, Y.; Wallach, D.; Liu, B.; Dingkuhn, M.; Boote, K.J.; Singh, U.; Asseng, S.; Kahveci, T.; He, J.; Zhang, R.; et al. Comparison of Three Calibration Methods for Modeling Rice Phenology. Agric. For. Meteorol. 2020, 280, 107785. [Google Scholar] [CrossRef]
  10. Thorp, K.R.; Drajat, D. Deep Machine Learning with Sentinel Satellite Data to Map Paddy Rice Production Stages across West Java, Indonesia. Remote Sens. Environ. 2021, 265, 112679. [Google Scholar] [CrossRef]
  11. Salsabila, C.; Ghazali, M.F.; Zaenudin, A. Historical Paddy Rice Growth and Phenology Pattern Estimation Using Dual Polarization of Sentinel 1. In Proceedings of the 2021 7th Asia-Pacific Conference on Synthetic Aperture Radar (APSAR), Kuta, Bali Island, Indonesia, 1–3 November 2021; pp. 1–5. [Google Scholar]
  12. Dey, S.; Bhogapurapu, N.; Bhattacharya, A.; Mandal, D.; Lopez-Sanchez, J.M.; McNairn, H.; Frery, A.C. Rice Phenology Mapping Using Novel Target Characterization Parameters from Polarimetric SAR Data. Int. J. Remote Sens. 2021, 42, 5515–5539. [Google Scholar] [CrossRef]
  13. Yang, H.; Pan, B.; Li, N.; Wang, W.; Zhang, J.; Zhang, X. A Systematic Method for Spatio-Temporal Phenology Estimation of Paddy Rice Using Time Series Sentinel-1 Images. Remote Sens. Environ. 2021, 259, 112394. [Google Scholar] [CrossRef]
  14. Chew, R.; Rineer, J.; Beach, R.; O’Neil, M.; Ujeneza, N.; Lapidus, D.; Miano, T.; Hegarty-Craver, M.; Polly, J.; Temple, D.S. Deep Neural Networks and Transfer Learning for Food Crop Identification in UAV Images. Drones 2020, 4, 7. [Google Scholar] [CrossRef] [Green Version]
  15. Colorado, J.D.; Calderon, F.; Mendez, D.; Petro, E.; Rojas, J.P.; Correa, E.S.; Mondragon, I.F.; Rebolledo, M.C.; Jaramillo-Botero, A. A Novel NIR-Image Segmentation Method for the Precise Estimation of above-Ground Biomass in Rice Crops. PLoS ONE 2020, 15, e0239591. [Google Scholar] [CrossRef]
  16. ten Harkel, J.; Bartholomeus, H.; Kooistra, L. Biomass and Crop Height Estimation of Different Crops Using UAV-Based Lidar. Remote Sens. 2019, 12, 17. [Google Scholar] [CrossRef] [Green Version]
  17. Lu, N.; Wu, Y.; Zheng, H.; Yao, X.; Zhu, Y.; Cao, W.; Cheng, T. An Assessment of Multi-View Spectral Information from UAV-Based Color-Infrared Images for Improved Estimation of Nitrogen Nutrition Status in Winter Wheat. Precis. Agric. 2022, 23, 1653–1674. [Google Scholar] [CrossRef]
  18. Yang, M.-D.; Tseng, H.-H.; Hsu, Y.-C.; Yang, C.-Y.; Lai, M.-H.; Wu, D.-H. A UAV Open Dataset of Rice Paddies for Deep Learning Practice. Remote Sens. 2021, 13, 1358. [Google Scholar] [CrossRef]
  19. Liao, K.-C.; Lu, J.-H. Using UAV to Detect Solar Module Fault Conditions of a Solar Power Farm with IR and Visual Image Analysis. Appl. Sci. 2021, 11, 1835. [Google Scholar] [CrossRef]
  20. Ge, H.; Ma, F.; Li, Z.; Tan, Z.; Du, C. Improved Accuracy of Phenological Detection in Rice Breeding by Using Ensemble Models of Machine Learning Based on UAV-RGB Imagery. Remote Sens. 2021, 13, 2678. [Google Scholar] [CrossRef]
  21. Ma, Y.; Jiang, Q.; Wu, X.; Zhu, R.; Gong, Y.; Peng, Y.; Duan, B.; Fang, S. Monitoring Hybrid Rice Phenology at Initial Heading Stage Based on Low-Altitude Remote Sensing Data. Remote Sens. 2021, 13, 86. [Google Scholar] [CrossRef]
  22. Yang, Q.; Shi, L.; Han, J.; Chen, Z.; Yu, J. A VI-Based Phenology Adaptation Approach for Rice Crop Monitoring Using UAV Multispectral Images. Field Crops Res. 2022, 277, 108419. [Google Scholar] [CrossRef]
  23. Qiu, Z.; Xiang, H.; Ma, F.; Du, C. Qualifications of Rice Growth Indicators Optimized at Different Growth Stages Using Unmanned Aerial Vehicle Digital Imagery. Remote Sens. 2020, 12, 3228. [Google Scholar] [CrossRef]
  24. Muharam, F.M.; Nurulhuda, K.; Zulkafli, Z.; Tarmizi, M.A.; Abdullah, A.N.H.; Che Hashim, M.F.; Mohd Zad, S.N.; Radhwane, D.; Ismail, M.R. UAV- and Random-Forest-AdaBoost (RFA)-Based Estimation of Rice Plant Traits. Agronomy 2021, 11, 915. [Google Scholar] [CrossRef]
  25. Yang, Q.; Shi, L.; Han, J.; Yu, J.; Huang, K. A near Real-Time Deep Learning Approach for Detecting Rice Phenology Based on UAV Images. Agric. For. Meteorol. 2020, 287, 107938. [Google Scholar] [CrossRef]
  26. Jin, B.; Cruz, L.; Gonçalves, N. Pseudo RGB-D Face Recognition. IEEE Sens. J. 2022, 22, 21780–21794. [Google Scholar] [CrossRef]
  27. Zheng, Q.; Yang, M.; Yang, J.; Zhang, Q.; Zhang, X. Improvement of Generalization Ability of Deep CNN via Implicit Regularization in Two-Stage Training Process. IEEE Access 2018, 6, 15844–15869. [Google Scholar] [CrossRef]
  28. Yao, T.; Qu, C.; Liu, Q.; Deng, R.; Tian, Y.; Xu, J.; Jha, A.; Bao, S.; Zhao, M.; Fogo, A.B.; et al. Compound Figure Separation of Biomedical Images with Side Loss. In Proceedings of the Deep Generative Models, and Data Augmentation, Labelling, and Imperfections; Engelhardt, S., Oksuz, I., Zhu, D., Yuan, Y., Mukhopadhyay, A., Heller, N., Huang, S.X., Nguyen, H., Sznitman, R., Xue, Y., Eds.; Springer International Publishing: Cham, Switzerland, 2021; pp. 173–183. [Google Scholar]
  29. Faster Mean-Shift: GPU-Accelerated Clustering for Cosine Embedding-Based Cell Segmentation and Tracking|Elsevier Enhanced Reader. Available online: https://reader.elsevier.com/reader/sd/pii/S1361841521000943?token=59BC9ED87AA5C3ED925A392D47B3B2CB2F49003C6976BD1201BBFCDE2A2E1D596A9A2DA8AC2B08377324479C0C036932&originRegion=us-east-1&originCreation=20230115150646 (accessed on 15 January 2023).
  30. Deng, J.; Zhong, Z.; Huang, H.; Lan, Y.; Han, Y.; Zhang, Y. Lightweight Semantic Segmentation Network for Real-Time Weed Mapping Using Unmanned Aerial Vehicles. Appl. Sci. 2020, 10, 7132. [Google Scholar] [CrossRef]
  31. Sai, G.U.; Tejasri, N.; Kumar, A.; Rajalakshmi, P. Deep Learning Based Overcomplete Representations for Paddy Rice Crop and Weed Segmentation. In Proceedings of the IGARSS 2022—2022 IEEE International Geoscience and Remote Sensing Symposium, Kuala Lumpur, Malaysia, 17–22 July 2022; pp. 6077–6080. [Google Scholar]
  32. Deng, H.; Yang, Y.; Liu, Z.; Liu, X.; Huang, D.; Liu, M.; Chen, X. A Paddy Field Segmentation Method Combining Attention Mechanism and Adaptive Feature Fusion. Appl. Eng. Agric. 2022, 38, 421–434. [Google Scholar] [CrossRef]
  33. Yang, M.-D.; Tseng, H.-H.; Hsu, Y.-C.; Tsai, H.P. Semantic Segmentation Using Deep Learning with Vegetation Indices for Rice Lodging Identification in Multi-Date UAV Visible Images. Remote Sens. 2020, 12, 633. [Google Scholar] [CrossRef] [Green Version]
  34. DJI-Zenmuse P1. Available online: https://www.dji.com/zenmuse-p1/specs (accessed on 12 December 2022).
  35. PIX4Dmapper-Support. Available online: https://support.pix4d.com/hc/en-us/categories/360001503192-PIX4Dmapper (accessed on 13 December 2022).
  36. QGIS: A Free and Open Source Geographic Information System. Available online: https://www.qgis.org/en/site/ (accessed on 13 December 2022).
  37. Hao, Y.; Liu, Y.; Wu, Z.; Han, L.; Chen, Y.; Chen, G.; Chu, L.; Tang, S.; Yu, Z.; Chen, Z.; et al. EdgeFlow: Achieving Practical Interactive Segmentation with Edge-Guided Flow. arXiv 2021, arXiv:210909406. [Google Scholar]
  38. OpenMMLab Semantic Segmentation Toolbox and Benchmark. Available online: https://github.com/open-mmlab/mmsegmentation (accessed on 13 December 2022).
  39. DJI-Matrice 300 RTK. Available online: https://www.dji.com/matrice-300/specs (accessed on 12 December 2022).
  40. Johnson, J.M.; Khoshgoftaar, T.M. Survey on Deep Learning with Class Imbalance. J. Big Data 2019, 6, 27. [Google Scholar] [CrossRef] [Green Version]
  41. Yu, C.; Wang, J.; Peng, C.; Gao, C.; Yu, G.; Sang, N. BiSeNet: Bilateral Segmentation Network for Real-Time Semantic Segmentation. In Proceedings of the European conference on computer vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 325–341. [Google Scholar]
  42. Yu, C.; Gao, C.; Wang, J.; Yu, G.; Shen, C.; Sang, N. BiSeNet V2: Bilateral Network with Guided Aggregation for Real-Time Semantic Segmentation. Int. J. Comput. Vis. 2021, 129, 3051–3068. [Google Scholar] [CrossRef]
  43. Lu, X.; Yang, R.; Zhou, J.; Jiao, J.; Liu, F.; Liu, Y.; Su, B.; Gu, P. A Hybrid Model of Ghost-Convolution Enlightened Transformer for Effective Diagnosis of Grape Leaf Disease and Pest. J. King Saud Univ.-Comput. Inf. Sci. 2022, 34, 1755–1767. [Google Scholar] [CrossRef]
  44. Han, K.; Wang, Y.; Tian, Q.; Guo, J.; Xu, C.; Xu, C. GhostNet: More Features From Cheap Operations. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, Seattle, WA, USA, 13–19 June 2020; pp. 1580–1589. [Google Scholar]
  45. Long, J.; Shelhamer, E.; Darrell, T. Fully Convolutional Networks for Semantic Segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  46. He, K.; Zhang, X.; Ren, S.; Sun, J. Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 1026–1034. [Google Scholar]
  47. Linder, W. Digital Photogrammetry; Springer: Berlin/Heidelberg, Germany, 2009; ISBN 978-3-540-92724-2. [Google Scholar]
  48. Wolf, P.R.; Dewitt, B.A.; Wilkinson, B.E. Elements of Photogrammetry with Applications in GIS, 4th ed.; [fully updated]; McGraw-Hill Education: New York, NY, USA, 2014; ISBN 978-0-07-176112-3. [Google Scholar]
  49. Flight Control-DJI Mobile SDK Documentation. Available online: https://developer.dji.com/mobile-sdk/documentation/introduction/flightController_concepts.html (accessed on 15 December 2022).
  50. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet Classification with Deep Convolutional Neural Networks. In Proceedings of the Advances in Neural Information Processing Systems, Lake Tahoe, NV, USA, 3–6 December 2012; p. 25. [Google Scholar]
  51. Hu, P.; Guo, W.; Chapman, S.C.; Guo, Y.; Zheng, B. Pixel Size of Aerial Imagery Constrains the Applications of Unmanned Aerial Vehicle in Crop Breeding. ISPRS J. Photogramm. Remote Sens. 2019, 154, 1–9. [Google Scholar] [CrossRef]
  52. Chan, K.C.K.; Zhou, S.; Xu, X.; Loy, C.C. BasicVSR++: Improving Video Super-Resolution with Enhanced Propagation and Alignment. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 19–24 June 2022; pp. 5962–5971. [Google Scholar]
  53. Hai, J.; Hao, Y.; Zou, F.; Lin, F.; Han, S. A Visual Navigation System for UAV under Diverse Illumination Conditions. Appl. Artif. Intell. 2021, 35, 1529–1549. [Google Scholar] [CrossRef]
  54. Hu, Z.; Shi, T.; Wang, C.; Li, Q.; Wu, G. Scale-Sets Image Classification with Hierarchical Sample Enriching and Automatic Scale Selection. Int. J. Appl. Earth Obs. Geoinf. 2021, 105, 102605. [Google Scholar] [CrossRef]
Figure 1. Experimental Sites Location and Orthophotos of Several Sites.
Figure 1. Experimental Sites Location and Orthophotos of Several Sites.
Drones 07 00083 g001
Figure 2. The construction process of PaddySeg dataset.
Figure 2. The construction process of PaddySeg dataset.
Drones 07 00083 g002
Figure 3. Similar Feature Maps in Convolution Neural Networks.
Figure 3. Similar Feature Maps in Convolution Neural Networks.
Drones 07 00083 g003
Figure 4. Structure illustration of (a) Ghost Convolution Module (GCM): initial feature maps are generated by C o n v 3 × 3 and expanded through G l i n e a r   ; (b) Ghost Block {stride 2-1}: is combined with 2 G C M with stride = 2 and 1 in order; (c) Ghost Convolution Network Segmentation Head (GCN-Head).
Figure 4. Structure illustration of (a) Ghost Convolution Module (GCM): initial feature maps are generated by C o n v 3 × 3 and expanded through G l i n e a r   ; (b) Ghost Block {stride 2-1}: is combined with 2 G C M with stride = 2 and 1 in order; (c) Ghost Convolution Network Segmentation Head (GCN-Head).
Drones 07 00083 g004
Figure 5. Overview of Ghost Bilateral Network (GBiNet). The network has 4 main parts: the Detail Branch, Semantic Branch, Aggregation Layer, and GCN Segmentation Head. Each box is an operation block, and the arrow connection represents the feature maps flow with numbers above showing the ratios of map size to the input size. The detail of each stage is available in Table 3.
Figure 5. Overview of Ghost Bilateral Network (GBiNet). The network has 4 main parts: the Detail Branch, Semantic Branch, Aggregation Layer, and GCN Segmentation Head. Each box is an operation block, and the arrow connection represents the feature maps flow with numbers above showing the ratios of map size to the input size. The detail of each stage is available in Table 3.
Drones 07 00083 g005
Figure 6. Direct Geo-Locating Based on (a) Exterior Orientation and (b) Central Projection at Nadir View.
Figure 6. Direct Geo-Locating Based on (a) Exterior Orientation and (b) Central Projection at Nadir View.
Drones 07 00083 g006
Figure 7. Evaluation of Direct Geo-Locating with (a) RTK and (b) Ground Control Points. Measuring the distance between the observed GCP location and the predicted location (marked with a blue dot) in one UAV image (c): Original Image; (d): Detail View of GCP.
Figure 7. Evaluation of Direct Geo-Locating with (a) RTK and (b) Ground Control Points. Measuring the distance between the observed GCP location and the predicted location (marked with a blue dot) in one UAV image (c): Original Image; (d): Detail View of GCP.
Drones 07 00083 g007
Figure 8. Incremental Sparse Sampling and System Workflow.
Figure 8. Incremental Sparse Sampling and System Workflow.
Drones 07 00083 g008
Figure 9. Convergence process of the models: (a). Loss descending during training; (b). Average segmentation accuracy on the validation set.
Figure 9. Convergence process of the models: (a). Loss descending during training; (b). Average segmentation accuracy on the validation set.
Drones 07 00083 g009
Figure 10. Segmentation Result of Rice Field at Different Stages.
Figure 10. Segmentation Result of Rice Field at Different Stages.
Drones 07 00083 g010
Figure 11. GSD Distance Deviation Distribution.
Figure 11. GSD Distance Deviation Distribution.
Drones 07 00083 g011
Figure 12. Rice Phenology Mapping at DJD-5 Experimental Field: (a). Drone Waypoints of Image Capture; (b). Sparse Sampled Patch-Boxes Distribution; (c). Distribution Map of Rice Phenology with Interpolation.
Figure 12. Rice Phenology Mapping at DJD-5 Experimental Field: (a). Drone Waypoints of Image Capture; (b). Sparse Sampled Patch-Boxes Distribution; (c). Distribution Map of Rice Phenology with Interpolation.
Drones 07 00083 g012
Figure 13. Detail View of the Singular Region in Phenology Map.
Figure 13. Detail View of the Singular Region in Phenology Map.
Drones 07 00083 g013
Figure 14. Normalized Confusion Matrix on Multi-Classes.
Figure 14. Normalized Confusion Matrix on Multi-Classes.
Drones 07 00083 g014
Table 1. Samples Number and Stages Distribution.
Table 1. Samples Number and Stages Distribution.
Mission
Code
Patch
Number
Class Pixel Number (Million)
SeedlingJointingHeadingFillingOthers
210604_djd1200208.8000149.1
210606_djd2140162.000088.5
210909_djd32000244.400113.4
210616_djd42000259.50098.4
210718_djd530034.3169.00186.2147.3
210721_djd630055.0184.50146.4150.8
220628_qt130087.36.5281.42.5159.0
220628_qt226076.073.4140.90174.9
220712_lq1100002.0117.759.2
220712_lq2100000119.459.5
220713_ra1100000119.159.8
220713_ra2100000122.756.2
220727_sds1400168.60081.9
220928_xs216002.8176.56.8100.2
Sum2600623.41108.7600.8820.81498.2
Pixel Ratio-13%24%13%18%32%
Table 2. Computation and Memory consumption of BiSeNetV2 [42].
Table 2. Computation and Memory consumption of BiSeNetV2 [42].
Computation/
Parameters
OverallComponents
Detail
Branch
Semantic
Branch
BGA
Layer
Decode
Head
Auxiliary
Head
FLOPs (G)21.2910.111.221.538.430.00
100.047.55.77.239.6/
Weights (M)3.34 10.521.160.481.1911.42
100.015.534.714.335.4/
1 Auxiliary Head weight is not included in the total for the inferencing process.
Table 3. Detailed Parameters of Standard GBiNet.
Table 3. Detailed Parameters of Standard GBiNet.
StageInput ShapeOperatorNumber and StrideOutput Shape
Encoder—Detail Branch 1H × W × 3G-Block 12-1H/2 × W/2 × 64
Encoder—Detail Branch 2H/2 × W/2 × 64G-Block 12-1-1H/4 × W/4 × 64
Encoder—Detail Branch 3H/4 × W/4 × 64G-Block 12-1-1H/8 × W/8 × 128
Encoder—Semantic Branch 1H × W × 3Stem-Block 34H/4 × W/4 × 16
Encoder—Semantic Branch 3H/4 × W/4 × 16GE-Block 32-1H/8 × W/8 × 32
Encoder—Semantic Branch 4H/8 × W/8 × 32GE-Block 32-1H/16 × W/16 × 64
Encoder—Semantic Branch 5H/16 × W/16 × 64GE-Block 32-1-1-1H/32 × W/32 × 128
Encoder—Semantic Branch 5H/16 × W/16 × 64CE-Block 31H/32 × W/32 × 128
Encoder—Aggregation Layer(H/8 × W/8 +
H/32 × W/32) × 128
BGA-Block 31H/8 × W/8 × 128
Decoder—Segmentation HeadH/8 × W/8 × 128GCN-Head 21H × W × 5
1 G-Block denotes Ghost Convolution Block. 2 GCN-Head means Ghost Convolution Networks for Semantic Segmentation. 3 Stem, GE, CE, and BGA-Block indicate the stem block, gather–expansion block, context embedding, and bilateral guided aggregation layer referring to BiSeNetV2.
Table 4. The Performance of Existing Models on PaddySeg (Slide Input: 546 × 546, Speed Tests on RTX2070).
Table 4. The Performance of Existing Models on PaddySeg (Slide Input: 546 × 546, Speed Tests on RTX2070).
ModelWeights
(M)
FLOPs
(G)
Speed
(FPS)
mIoU-val
%
aAcc-Test
%
mIoU-Test
%
DSW Speed (FPS)
pspnet_r18-d8 (2017)12.7962.982.891.7895.5391.5721.3
deeplabv3p_r18-d8 (2018)12.4762.912.792.1695.5691.5520.7
fcn_hr18s (2019)3.9411.132.591.8295.3991.2821.7
bisenetv2_fcn (2021)14.7714.255.491.5295.4191.3136.4
Table 5. The Performance of GBiNet on PaddySeg (DSW input: 819 × 546, Speed Tests on RTX2070).
Table 5. The Performance of GBiNet on PaddySeg (DSW input: 819 × 546, Speed Tests on RTX2070).
ModelWeights (M)FLOPs (G)DSW Speed (FPS)mIoU-valaAcc-TestmIoU-Test
bisenetv2_fcn14.7721.2936.491.2595.1190.95
GBiNet_r213.9312.2241.091.6495.4391.50
GBiNet_r413.517.6844.991.1394.8990.47
GBiNet_r813.305.4146.890.9294.9390.56
GBiNet_64dx4_r23.513.0347.990.7494.7990.26
GBiNet_64dx8_r43.342.2452.390.9094.8090.40
GBiNet_t32dx2_r40.820.5061.990.2094.7190.19
Table 6. Computation Costs and Parameters Number of GBiNet and BiSeNetV2.
Table 6. Computation Costs and Parameters Number of GBiNet and BiSeNetV2.
Computation/
Parameters
ModelOverallComponents
Detail
Branch
Semantic
Branch
BGA
Layer
Decode
Head
Auxiliary
Head
FLOPs (G)bisenetv2_fcn21.28810.1131.2231.5258.4270
GBiNet_r212.2225.1911.2231.5254.2830
GBiNet_r47.6812.7301.2231.5252.2030
GBiNet_r85.4121.5001.2231.5251.1640
GBiNet_64dx8_r42.4430.7490.7310.3850.5780
GBiNet_t32dx2_r40.4990.1820.1800.0980.0390
Weights (M)bisenetv2_fcn3.3430.5191.1600.4791.18511.421
GBiNet_r22.5040.2631.1600.4790.60211.421
GBiNet_r42.0840.1361.1600.4790.30911.421
GBiNet_r81.8740.0721.1600.4790.16311.421
GBiNet_64dx8_r40.5770.0360.3390.1210.0812.902
GBiNet_t32dx2_r40.0910.0060.0490.0310.0050.726
Table 7. Segmentation Results over Multi-Classes on PaddySeg Test Set.
Table 7. Segmentation Results over Multi-Classes on PaddySeg Test Set.
ModelClass Pixel Number Ratio and IoU-Class (%)
SeedlingJointingHeadingFillingOthers
13%24%13%18%32%
bisenetv2_fcn91.7924%88.2094.3887.54
GBiNet_r291.6293.4389.8594.3188.29
GBiNet_r491.2992.2187.1894.2187.48
GBiNet_r890.4992.5988.6793.5887.46
GBiNet_64dx8_r489.9891.9089.3993.7487.01
GBiNet_64dx4_r290.4291.8287.7293.8987.45
GBiNet_t32dx2_r488.8392.3189.9093.3186.61
Average90.6379.2188.7093.9287.41
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lu, X.; Zhou, J.; Yang, R.; Yan, Z.; Lin, Y.; Jiao, J.; Liu, F. Automated Rice Phenology Stage Mapping Using UAV Images and Deep Learning. Drones 2023, 7, 83. https://doi.org/10.3390/drones7020083

AMA Style

Lu X, Zhou J, Yang R, Yan Z, Lin Y, Jiao J, Liu F. Automated Rice Phenology Stage Mapping Using UAV Images and Deep Learning. Drones. 2023; 7(2):83. https://doi.org/10.3390/drones7020083

Chicago/Turabian Style

Lu, Xiangyu, Jun Zhou, Rui Yang, Zhiyan Yan, Yiyuan Lin, Jie Jiao, and Fei Liu. 2023. "Automated Rice Phenology Stage Mapping Using UAV Images and Deep Learning" Drones 7, no. 2: 83. https://doi.org/10.3390/drones7020083

APA Style

Lu, X., Zhou, J., Yang, R., Yan, Z., Lin, Y., Jiao, J., & Liu, F. (2023). Automated Rice Phenology Stage Mapping Using UAV Images and Deep Learning. Drones, 7(2), 83. https://doi.org/10.3390/drones7020083

Article Metrics

Back to TopTop