Remote Sensing Identiﬁcation and Rapid Yield Estimation of Pitaya Plants in Diﬀerent Karst Mountainous Complex Habitats

: The Pitaya industry is a specialty fruit industry in the mountainous region of Guizhou, China. The planted area in Guizhou reaches 7200 ha, ranking ﬁrst in the country. At present, Pitaya planting lacks eﬃcient yield estimation methods, which has a negative impact on the Pitaya downstream industry chain, stymying the constant growing market. The fragmented and complex terrain in karst mountainous areas and the capricious local weather have hindered accurate crop identiﬁcation using traditional satellite remote sensing methods, and there is currently li�le a�empt made to tackle the mountainous specialty crops’ yield estimation. In this paper, based on UAV (unmanned aerial vehicle) remote sensing images, the complexity of Pitaya planting sites in the karst background has been divided into three diﬀerent scenes as complex scenes with similar colors, with topographic variations, and with the coexistence of multiple crops. In scenes with similar colors, using the Close Color Vegetation Index (CCVI) to extract Pitaya plants, the accuracy reached 92.37% on average in the sample sites; in scenes with complex topographic variations, using point clouds data based on the Canopy Height Model (CHM) to extract Pitaya plants, the accuracy reached 89.09%; and in scenes with the coexistence of multiple crops, using the U-Net Deep Learning Model (DLM) to identify Pitaya plants, the accuracy reached 92.76%. Thereafter, the Pitaya yield estimation model was constructed based on the fruit yield data measured in the ﬁeld for several periods, and the fast yield estimations were carried out and examined for three application scenes. The results showed that the average accuracy of yield estimation in complex scenes with similar colors was 91.25%, the average accuracy of yield estimation in scenes with topographic variations was 93.40%, and the accuracy of yield estimation in scenes with the coexistence of multiple crops was 95.18%. The overall yield estimation results show a high accuracy. The experimental results show that it is feasible to use UAV remote sensing images to identify and rapidly estimate the characteristic crops in the complex karst habitat, which can also provide scientiﬁc reference for the rapid yield estimation of other crops in mountainous regions.


Introduction
The karst regions account for about 12% of the global surface [1]. Guizhou is located in the central area of the karst region of southwest China, one of the three largest karst regions in the world [2][3][4][5]. In total, 92.5% of the province is covered by mountains and hills [3]. The complex and diverse ecological environment of Guizhou is rich in agricultural resources. Under the influence of both climate change and human activities, the spatial and temporal differences in crop cultivation structure in karst regions are significant [6,7]. Rapid investigation of a crop cultivation area, efficient identification of crop types, and accurate estimation of crop production value are prerequisites for highquality monitoring and scientific management of agricultural sites in mountainous areas. The deep integration of karst mountain agriculture development with multi-source remote sensing technology and spatial information technology is of great significance to grasp the current situation of mountain agriculture planting in a timely and accurate manner and improve the level and precision of mountain agriculture planting monitoring.
As the advantageous industry in the Guizhou mountainous region, the specialty fruit industry has played an important role in the construction of ecological civilization and helped the increase in local farmers' income and the integrated development of the major industries. In 2020, China's Pitaya planting area reached more than 20,000 ha, while the Guizhou planting area reached 7200 ha, ranking first in the country [8]. The market capacity of Pitaya is estimated to grow to 2-3 million tons in the next 5-10 years [9].
To be er suit the constantly growing market, the purpose of implementing rapid yield estimation of Pitaya is to be er manage and plan Pitaya planting activities. By accurately estimating yields, local farmers can be er control costs, arrange sales plans, and be er adapt to market demand. In addition, rapid yield estimates can also help growers take timely measures to increase yield and improve the production process. The Chinese Pitaya planting area is extensive, and the rapid estimation of Pitaya is very important to improve the yield efficiency and quality of Pitaya growers so as to enhance their competitiveness and market position.
Despite being an important part of the specialty fruit industry in Guizhou province, local Pitaya plantations are still in a more traditional stage. Pitaya plant monitoring and yield estimation still largely rely on manual execution, with low efficiency and high cost expenditure, which present a negative impact on the overall development of the Pitaya industry. However, intelligent remote sensing monitoring and yield estimation of Pitaya are able to provide a reference for scientific planning of the Pitaya industry's layout and for industrial upgrading and transformation. Remote sensing yield estimation is based on remote sensing technology and digital image processing technology, which are able to quickly and accurately obtain the information of large areas of farmland and realize the accurate measurement of crop types, growth conditions, and the coverage degree of different farmlands. Therefore, compared with the traditional yield estimation method, remote sensing yield estimation has the advantages of high efficiency, low cost, and high accuracy, and it is a more advanced and reliable production estimation method. At present, there is no Pitaya estimation method based on remote sensing technology, so it is particularly necessary to develop a Pitaya estimation method based on remote sensing.
In recent years, remote sensing technology has reached the practical stage, and it has been widely used in crop yield estimation with its advantages of wide coverage, fast updating speed, and non-destructive detection [10]. Liu et al. (2022) proposed a novel deep learning adaptive crop model (DACM) to accomplish adaptive high-precision yield estimation in large areas, which emphasized adaptive learning of the spatial heterogeneity of crop growth based on fully extracting crop growth information; the results showed that the DACM achieved an average root mean squared error (RMSE) of 4.406 bushels·acre −1 (296.304 kg·ha −1 ), with an average coefficient of determination (R 2 ) of 0.805 [11].  developed an NPP-based rice yield estimation model, with HI (harvest index) being a changing parameter, and the accuracy of the proposed method was tested over the Jiangsu Province, southeast China, which showed that the rice HI increased linearly in the study area. Compared with yield estimation with a fixed harvest index, yield estimations using a changing HI were greatly improved [12]. He et al. (2023) proposed a novel yield evaluation method for integrating multiple spectral indexes and temporal aggregation data, and the results showed that the normalized moisture difference index (NMDI) was the index most sensitive to yield estimation [13]. Liu et al. (2016) established a remote sensing yield estimation model for complex mountainous areas based on ZY-3 satellite remote sensing images and concluded that the model constructed by NDVI had the best fit with the actual yield [14]. These studies mainly focused on vegetation indices and spectral data obtained by remote sensing technology for large-scale crop yield estimation; the most common observation method used NDVI and EVI indices to estimate biomass, crop yield, and other factors to reflect crop growth characteristics. However, former studies have usually focused on concentrated bulk crops grown in large-scale farmland. There are currently few studies on high-precision plant extraction and highprecision yield estimation for crops with large planting gaps, such as Pitaya.
Traditional satellite remote sensing yield estimation is applicable to large-scale cropland. However, crops in karst mountainous complex habitats are small and sca ered, and they are easily affected by bad weather (e.g., cloudy and rainy), so that satellite remote sensing data cannot be obtained as accurate ground information in time or effectively; thus, the traditional satellite remote sensing method is generally ineffective.
The rapid development of UAV remote sensing technology has avoided the shortages of satellite remote sensing yield estimation and can be applied to small-scale cropland [15]. UAV remote sensing estimation mainly imitates satellite remote sensing estimation methods, but it has the advantages of low cost, high spatial resolution, real time, and low atmospheric disturbance, which provide a new technical means for remote sensing crop estimation on a small scale [16][17][18][19][20]. For example, Chen et al. (2019) have compared humanengineered features (namely, histogram of oriented gradients (HOG), local binary pa ern (LBP), and scale-invariant feature transform (SIFT)) with features extracted using a pretrained Alex-net model for detecting palm trees in high-resolution images obtained by UAV [21]. Sothe et al. (2020) obtained the classification of 14 tree species in a subtropical forest area of Southern Brazil by UAV data [22]. Al-Ruzouq et al. (2023) proposed a fixedwing UAV system equipped with an optical (red-green-blue) camera used to capture very high spatial resolution images over urban and agricultural areas in an arid environment [23]. Wang (2015) referenced the form of NDVI and put forward a new vegetation index, VDVI (visible-band difference vegetation index), by using UAV data; the total extraction precision was 91.5%, and the kappa coefficient was 0.8256 [24]. From the above studies, it can be found that there are many studies on UAV image recognition, while there are fewer studies on further application of the recognition results. There are some studies that have discussed the identification and estimation of corps in a single scene, but there are few studies on identification and estimation in multiple scenes.
The development of agriculture in karst mountainous areas is highly limited by the topography, small habitat environment, and climate, and it is difficult for traditional satellites to acquire data that meet the accuracy requirements due to the influence of clouds, rain, and fog. Low-altitude UAV remote sensing has flexible periodicity and timelines for crop growth data collection and monitoring. In this paper, we used UAV as a means of remote sensing data collection, and we took Pitaya, a special but important crop of the karst mountains, as the research object. We established a knowledge base of remote sensing recognition features of Pitaya, designed remote sensing recognition algorithms and models of crop recognition based on deep learning, classified mountain agriculture according to the composition of complex backgrounds, and constructed technical methods for plant extraction and yield estimation in complex scenes of similar colors, terrain changes, and the coexistence of multiple types of crops. The technical method of yield estimation proposed in this study is able to improve the utilization rate of mountain agricultural resources information and provide decision support, scientific support, and technological support for the development of modern mountainous agriculture in the karst area.

Study Area
In this study, the total four sample sites are located in Huajiang town and Luodian County of southern Guizhou Province (Figure 1a,b). One of the sample sites is located in  The general topography of Luodian County is high in the north and low in the south, with a stepped descent, and the elevation of the county is mostly between 300 and 1000 m. Typical karst mountains, gullies, and ravines account for 85.75% of the total area, and hills and basins are distributed among each other [25]. The climate in this area is a subtropical monsoon climate, featuring an early spring, a long summer, a late autumn, and a short winter, with 1350-1520 h of sunshine, an average annual temperature of 20 °C, an average annual rainfall of 1335 mm, and a frost-free period of 335 days. The rivers in the county belong to the Red River system of the Pearl River Basin, with a total length of 482 km and an average annual runoff of 1.778 × 109 m3. The soil is mostly sandy red soil and red brown soil [26]. Luodian County is known as the "natural greenhouse," and it is the ideal place to plant Pitaya.
The Huajiang demonstration area is of typical karst landscape, too [27]. The climate is hot and dry, with an annual rainfall of 1120 mm; May to October is the frequent period of precipitation, accounting for 83% of the total annual precipitation. The average annual temperature is 18.4 °C, and the soil type is mainly brownish yellow lime soil and black lime soil. The high mountains and steep slopes in the territory cause frequent soil erosion, resulting in infertile land and serious stone desertification [28]. In order to alleviate the intensification of stone desertification, and, at the same time, to bring economic benefits, the local government vigorously promoted the development of the Pitaya fruit industry according to local conditions [29].

Materials
(a) UAV remote sensing data. Remote sensing data were acquired using DJI Mavic 2 Pro (Manufactured by DJ Innovations, Shenzhen, China) platforms carrying visible lens loads with 20 million effective sensor pixels. The remote sensing data were acquired on 15 July 2022, 13 August 2022, 9 September 2022, and 18 October 2022, respectively. The images of the Red Heart Pitaya Tourism Park were acquired on 14 July 2022, 14 August 2022, 7 September 2022, and 17 October 2022. The aerial video was taken during the summer fruiting period. The shooting time was 12:00-12:40 pm, with clear light, moderate wind, and sunlight close to vertical ground, which minimized crop shadows during the shooting process. The route was planned using Altizure-v3.9.5 software. The topography of the Pitaya planting area was obviously undulating, so the route and side overlap were set to 80%, and the flight height was 60 m (pixel size 1.48 cm × 1.48 cm) to ensure data validity ( Figure 2 and Table 1).    Figure 4a-d. From the processed images in the study area, we have divided out three similar color scenes, two complex topographic variation scenes, and one multiple crop coexistence scene, to implement the experiment.

Research Flow Chart
The technical route of the research is as follows: (1)

Remote Sensing Identification Method for Pitaya Plants in Complex Habitats
(1) CCVI (Close Color Vegetation Index). For similar color complex scenes, use CCVI [30] for vegetation extraction. Close color complex scenes are remote sensing application scenes that possess similar color interference, such as wild weeds. In order to highlight the color difference between plants and weeds, combined with the numerical relationship between them in R, G, and B reflectance, the CCVI of similar color vegetation index based on visible light red, green, and blue 3 bands is constructed based on the principle of color addition by plants to R and G reflectance, calculated as the following equation: where R, G, and B represent the red, green, and blue bands, respectively. The density partitioning of the gray space after CCVI extraction is performed, and the binarization calculation of the gray space of Pitaya plants and weeds is performed using the bipartite histogram and the maximum class variance threshold extraction method to obtain the binarized gray space containing the target features and background values; then, the Pitaya plants are extracted.
(2) CHM (Canopy Height Model). For complex topographic variation scenes, we use point cloud data to extract plants based on CHM (Canopy Height Model) [31,32]. CHM segmentation is a kind of raster image segmentation method, which mainly includes a watershed algorithm and a water injection algorithm. CHM segmentation removes noise from the original point cloud data and then performs a filtering process on the point cloud data, i.e., point cloud classification, and divides it into ground points and non-ground points. The ground point generates a digital elevation model (DEM) and the non-ground point generates a digital surface model (DSM), and the difference between them is used to calculate the Canopy Height Model (CHM). The generation of CHM is actually the establishment of the difference model between DSM and DEM to represent the height, location, and surface size of objects on the ground with the following equation: CHM-based plant extraction is a process of fixing the point location at the maximum value of local brightness and gradually decreasing along the edge direction. Combined with the actual measurement data of plant height of sample plants, the plant CHM is established, and the threshold range is set to obtain the CHM segmentation-based Pitaya plants.  In Figure 6, the bars indicate the feature map, and the top number represents the number of channels; the blue bars represent multichannel features; the gray bars represent duplicate multichannel features; the purple arrows indicate 3 × 3 convolution for feature extraction; the gray arrows indicate jump links for feature fusion; the red arrows indicate pooling for dimensionality reduction; the green arrows indicate up sampling for image size recovery; and the orange arrows indicate 1 × 1 convolution, which is used for outpu ing the results. Using the constructed Pitaya crop samples, the U-Net network model architecture learning samples are extracted to achieve efficient remote sensing recognition of Pitaya under complex scenes of multi-class crop coexistence based on U-Net deep learning.

Evaluation of the Accuracy of Pitaya Plant Identification
We use the detection rate (R), correct rate (P), and overall accuracy (F) to evaluate the accuracy of Pitaya plant identification.
(1) The detection rate (R) calculation formula is as Equation (3): where TP is the number of correctly identified plants and Nr is the actual number of plants.
(2) The accuracy rate (P) is calculated by Equation (4): where TP is the number of correctly identified plants and Nc is the number of identified plants.
(3) The overall precision (F) is calculated as Equation (5) where R is the detection rate and P is the accuracy rate.

Pitaya Yield Estimation Model
The average yield of single plants was used to estimate the yield of the featured crops in color similar scenes, terrain variation scenes, and coexistence of multiple crops scenes, respectively. The calculation method is shown in Equation (6): ) m where I is the yield, N is the number of individual plants of the crop, m1 is the average number of large fruits per plant, m2 is the average number of medium fruits per plant, m3 is the average number of small fruits per plant, Y1 is the average weight of large fruits, Y2 is the average weight of medium fruits, and Y3 is the average weight of small fruits.

CCVI Calculation
Using the band calculation tool of ENVI software [7,14,29,30], a remote sensing data processing platform, the similar color vegetation index model was used to calculate the CCVI (Close Color Vegetation Index) of the three sample plots (segment from Sunshine Orchard, Xinzhongsheng Pitaya Planting Base, and Langdang Pitaya planting base) containing Pitaya plants and weeds to obtain the color gray space, as shown in Figure 7a

CCVI Plant Identification
The grayscale spatial distribution of plants and weeds was obtained by grayscale density segmentation, and the maximum between-cluster variance method (OSTU method) [35] was used to binarize the grayscale of plants and weeds. Because the CCVI index was calculated in image elements, the binarization results showed some discontinuous fragmented patches and patch hollow phenomenon, and it used the processing methods of convolution filtering and cluster analysis to remove the fragmented patches and fill the patch hollow area. The three sample plots and the corresponding extraction results were obtained, as shown in Figure 8, for example.

Evaluation of the Accuracy of CCVI Plant Identification
Among the three sample areas, sample area 2 had the best extraction results, while sample area 1 and sample area 3 were affected by the topography, and the accuracy of digital orthophoto (DOM) was impaired, which caused the variation of DN values and resulted in a small number of missing extraction plants. On the other hand, because wild weeds and plants did not completely cover the whole sample area, there was some bare soil and a few field canyons, and the bare soil and field canyons were mistakenly treated as plants in the calculation of the index, which caused some over extraction errors. The actual number of Pitaya plants in the three sample areas, the CCVI index extracted plants, and the correct rate were obtained through human-machine interaction, respectively, as shown in Table 2. Comparing the actual number of plants and the number of extracted plants in the three sample plots, the over extraction 14 plants were extracted from Sunshine Orchard, the over extraction 5 plants were extracted from Xinzhongsheng, and 10 plants were missed from Langdang; the overall precision, detection rate, and accuracy rate were all greater than 90%, which met the general requirements of extraction precision.

Point Cloud Characteristics of Pitaya Plants
The CHM (Canopy Height Model) was used to identify Pitaya plants in complex terrain variation scenes of the two sample plots (segment from Sunshine Orchard and Guanling Pitaya planting demonstration area). Two separate plants, two plants with branches crossing each other, and plants next to wild weeds were selected as samples, and the point cloud morphology of Pitaya plants under different growth environments was extracted by combining visible image data. The results are shown in Figure 9.
In Figure 9a-c, the separate Pitaya plants each have independent structural features, and there is no mutual interference with each other, which is good for separation. In Figure 9d-f, the branches of two Pitaya plants cross each other, and there is confusion of point clouds when acquiring. In Figure 9g-i, the point cloud data of wild weeds and Pitaya are plo ed, and the point cloud data of Pitaya are obviously higher than those of weeds; the point cloud of weeds is close to the ground in space, which will not affect the extraction of Pitaya. In summary, the point cloud data distribution of Pitaya has an obvious spatial hierarchy, and the method of extracting single wood can be applied to the extraction of Pitaya plants.

Pitaya Recognition based on CHM Segmentation and Accuracy Analysis
A watershed algorithm for CHM segmentation [31,32] was used to consider the image of the Pitaya canopy surface as a topological landscape in geodesy, and the gray value of each pixel in the image represents the elevation of that point. Each local minimal value and its influence area is called the catchment basin, which is the Pitaya canopy, and the boundary of the catchment basin forms the watershed, which is the Pitaya canopy boundary (Figure 10).  The identified Pitaya plant canopy points were combined with the visually identified Pitaya plant location points to finally obtain their identification accuracy (Table 3).

Identification of Pitaya Plants in Complex Scenes with Multiple Crop Species Coexistence
In complex habitats with multiple crop types coexisting and large topographic undulations, resulting in the diminished significance of spectral feature differences, point clouds are difficult to distinguish crops. This study uses the U-Net deep learning network model to identify Pitaya plants in complex scenes with multiple crop species coexistence of the sample plots (segment from Langdang Pitaya planting base).

Recognition Sample Training and Analysis
Construction of the training sample set. In order to realize the differentiation of data, the differentiation model is obtained by training the U-Net deep learning network. The training sample set was established by combining samples of multiple landscapes and multiple influencing factors [36,37], and then the samples were binarized to establish a large number of sample sets with high richness and high accuracy (Figure 11). Sample set slicing. Due to the special structure of down sampling and up sampling of the U-Net network model, though it does not limit the size of the image, the sample set with too large a size will bring a negative impact on the network training time. The slicing of the sample set can be er reduce the training time and improve the model efficiency. During training, we divided the regional sample set into 224 × 224 sample slices and selected 100 slices containing Pitaya plants in the sample slices to construct the training sample data.
Sample set classification. The sample data were classified into the training set, the validation set, and the test set, with the proportion of the training set being 60%, the proportion of the validation set being 20%, and the proportion of the test set being 20%. The training set was used to fit the data samples for the model and to debug the parameters in the network; the validation set was used to see whether the effect of model training proceeds in a bad direction, which was reflected in the training process; and the test set was used to evaluate the generalization ability of the final model. The division of the original data into three datasets prevented the model from being overfi ed. The training process of Pitaya plant samples is shown in Figure 12.

Analysis of Recognition Results and Accuracy
Using the U-Net deep learning network model, recognition samples of multiple crops were constructed for deep learning recognition extraction, and the results of Pitaya plant extraction are shown in Figure 13. The segmentation results of the U-Net model of the sample were analyzed, and the accuracy of the detection rate, the correct rate, and the overall accuracy reached more than 90% (Table 4).

Yield Estimation Model Construction
Five batches of, in total, 800 sets of yield data were obtained by actual measurement, and the average weights of large, medium, and small fruits of a single plant were measured as 402.33 g, 274.73 g, and 203.12 g, respectively. The model of Pitaya yield estimation is: where I is the yield and N is the number of individual plants of the crop.

Yield Estimation in Different Habitats
(1) Fast estimation of yield in complex scenes of similar colors Under the conditions of complex scenes with similar colors, such as wild weed interference, the number of extracted plants in the three sample areas was obtained using CCVI. This was combined with the Pitaya estimation model to obtain the estimated yield results of the three sample sites, as shown in Table 5. (2) Rapid yield estimation under complex scenes of terrain variation Under the conditions of complex scenes with terrain variation, the CHM was used to obtain the number of extracted plants in the two sample sites and combined with the Pitaya estimation model to obtain the estimated yield results of the two sample sites, as shown in Table 6.   Table 7. From the results, it can be seen that the total precision of the estimated Pitaya yield results compared to the actual results of the sample site is 95.18%, which has a high degree of precision.

Discussion
The terrain is fragmented and complex in karst mountainous areas, causing farmlands in karst mountainous areas to become small and sca ered, which means they are not easily revealed on traditional satellite remote sensing pictures. The subtropical monsoon climate in Guizhou Province exacerbates the situation because traditional satellite remote sensing is easily affected by clouds and rain. So, because it is generally ineffective for crops identification in karst mountainous areas to use traditional satellite remote sensing methods, yield estimation is even harder to realize. In this article, we explored the use of UAV remote sensing to increase plants extraction precision, and proposed a new model for yield estimation of the special plant Pitaya. The accuracy of the methods promoted in this article show high accuracy, but their applicability is still an issue that deserves further exploration.

The Use of UAV Remote Sensing
Low-altitude UAV remote sensing has flexible periodicity and timelines for crop growth data collection and monitoring. Based on UAV remote sensing images, this study used multiple identification methods and remote sensing software to accurately identify individual Pitaya plants and successfully established a Pitaya yield estimation model based on fruit production data of sample plants measured in the field for multiple periods to perform rapid yield estimation for three application scenes. However, UAV remote sensing technology is still in a stage of continuous development, and high-quality image acquisition is affected by various factors [16]. For example, UAVs are limited by their ba ery capacity, and it may take several flights to complete aerial photography within one region; as such, it is necessary to consider the influence of flight time and spatial resolution, overlap se ings, load, etc. [38]. Reducing the flight altitude is one way to improve the spatial resolution of the image, but it could lead to a reduction in ground coverage and an increase in flight time. Increasing the degree of overlap can improve the image stitching effect, but it will also increase the flight time [39]. So, a reasonable flight mission needs to be tailored according to the research needs and field conditions.

Precision of Identifying Plants in the Images or Similar
In the complex terrain of karst mountains, multiple identification methods are used for accurate identification of Pitaya plants in different scenes. In scenes where Pitaya plants mix with wild plants of the same color, the average accuracy of Pitaya identification based on the CCVI (Close Color Vegetation Index) method reached 92.37%; in scenes of complex terrain variation, the average accuracy of Pitaya identification based on the CHM reached 89.09%; and in scenes where multiple crops coexist, the average total accuracy of Pitaya identification based on the U-Net deep learning model reached 92.76%. As the numbers indicate, the overall accuracy of the methods promoted in this article is high, hence proving the effectiveness and necessity of the application of different methods for complex karst region crop identification using remote sensing. Compared with previous studies, in which the identification of bulk crops is mainly distributed in plain areas [40][41][42], this study firstly targeted the broken and sca ered planting areas of characteristic crops in karst complex areas and originally implemented crop identification methods divided according to different background scenarios. By dividing the study area into small areas of different scenarios, we improved the scope limitation of crop identification research and fixed the inapplicability of normal methods caused by complex surface variation and fragmentation in karst areas. Compared with existing studies [30,31,[43][44][45][46][47][48][49][50][51], this broadens the scope of UAV in surface parameter inversion, too.

The Yield Estimation Model and Its Applicability
There are currently few studies on yield estimation of Pitaya using remote sensing, and there is no unified standard yield estimation model yet. In this study, we propose a new method using multi-period field-collected yield data in several sample areas to construct a yield estimation model. This model fills the gap in the yield estimation research of characteristic crops in karst complex areas and expands the further application of crop characteristics extracted by UAV, though the applicability of this model to areas other than karst terrain needs further verification. The next step of this study is to collect yield data from Pitaya plants in multiple regions, expand the test area, so as to explore regional differences in the yield estimation of Pitaya, and improve the applicability of the model. Secondly, in this study, the obtained yield estimation model was limited to the peak production period of Pitaya. Further consideration could be given to obtaining image data of multiple Pitaya production periods for many years and collecting yield data for the same period to explore a yield estimation model applicable to different Pitaya production periods, and then improving the accuracy of the Pitaya yield estimation model.

Conclusions
This study, based on UAV remote sensing data, uses various identification methods and remote sensing software to realize the identification of Pitaya plants with high accuracy in different complex scenes, and establishes a new Pitaya yield estimation model based on Pitaya yield. The main findings are as follows: (1) The application of different identification methods and remote sensing software can accurately identify individual Pitaya plant in complex scenes of karst background. The total precision is high, which demonstrates the effectiveness and necessity of using UAV remote sensing data to divide different scenes and of implementing multiple methods to realize the identification of specialty crops in complex karst areas. (2) Based on fruit production data of sample plants measured in the field and the identified Pitaya plants, a rapid yield estimation model for three different scenes was constructed. The yield estimation results show high accuracy, which implicates the effectiveness of the rapid yield estimation model constructed in this study. (3) This study has preliminarily proven that it is feasible to use UAV remote sensing images for single Pitaya plant identification and, therefore, to implement rapid yield estimation.