1. Introduction
In the precision production of orchards, scientific control of flower and fruit retention is the core link to achieve high quality and increased yield. Among them, as a critical component of orchard management, flower thinning operation plays an irreplaceable role in balancing tree nutrition distribution and improving fruit quality [
1]. Flat peach, as a kind of fruit with tough skin and easy peeling, delicate flesh, and sweet juice, is deeply adored by consumers [
2], and it is crucial to perform thinning operations on it. According to the literature investigation, flower thinning technology can be categorized into three types. These types include manual flower thinning, chemical flower thinning (which usually refers to the employment of chemical thinning agents for the operation), and mechanical flower thinning [
3]. While manual flower thinning ensures a certain level of operational precision, it is labor-intensive and costly, presenting several limitations in practical applications [
4]. Meanwhile, chemical flower thinning is susceptible to various factors such as climatic conditions, tree age, and tree vitality [
5]. Furthermore, chemical substances may persist after usage, which contradicts contemporary principles of pollution-free fruit and vegetable production. Previous pump-type mechanical flower-thinning equipment often inflicts unnecessary damage to trees while demonstrating suboptimal effectiveness in flower thinning [
6]. Therefore, exploring and implementing intelligent and precise flower thinning solutions that integrate machine vision with other artificial intelligence technology is particularly crucial [
7].
Review the existing literature, traditional machine vision technology primarily depends on key feature information such as color, shape, texture, and spectral information to detect flowers and fruits [
8,
9]. Among them, color features have been extensively utilized in flower recognition. For instance, Hočevar et al. [
10] employed the HSL color space and threshold method to precisely assess the flowering intensity of individual apple trees. Pornpanomchai et al. [
11] combined RGB color information with the size characteristics of herbal flowers and the complex characteristics of petal edges to develop an intelligent approach for the automatic identification of herbal flowers. Similarly, shape characteristics also play a role in the detection of flowers and fruits. Krikeb et al. [
12] combined morphological techniques with the threshold method, successfully identifying flowers while also predicting the specific blooming time period. Utilizing contour segments and color information, Lu et al. [
13] developed an effective method for the identification of citrus in various lighting conditions. Meanwhile, in the utilization of texture features for fruit identification, Lin et al. [
14] proposed a support vector machine (SVM) model that employs color and texture features to accurately detect a range of fruits, including citrus and tomatoes. In practical applications, Wouters et al. [
15] developed a multi-spectral camera system to establish an accurate discrimination model by collecting the pixel information of buds and to achieve precise prediction of the number of buds (with an accuracy of 87%). In the aforementioned research, the recognition technology employing the traditional image processing algorithm is prone to interference from factors such as changes in illumination, diversity of forms, complexity of textures, and complexity of the environmental background, and its applicability and stability are low [
16]. By contrast, in recent years, Deep Learning (DL) technology has demonstrated superior outcomes in the domain of orchard flower recognition.
By extracting more abundant and profound feature information from image data, deep learning technology significantly reduces the influence of surrounding environmental factors on image processing results. And it effectively enhances the precision of object detection in complex dynamic environments. Owing to the powerful feature extraction and classification capability of deep learning, semantic segmentation performs well in flower recognition. For instance, Dias et al. [
17] and Sun et al. [
18] employed the DeepLab+RGR and DeepLab-Resnet semantic segmentation models, respectively, to detect apple flowers, peach flowers, and pear flowers, and both achieved highly precise segmentation of the target flowers. Concurrently, Wang et al. [
19] introduced a method based on Fully Convolutional Networks (FCN) for precise pixel-level segmentation of apple flowers (with an average KL value of 0.23). Furthermore, to meet the purpose of precise flower thinning and to identify the same type of flowers at different growth stages, the case segmentation technique is employed. Tian et al. [
20] proposed a mask scoring R-CNN instance segmentation model (MASU R-CNN) utilizing a U-Net backbone network, which was specifically optimized for three different growth stages of apple flower (bud stage, semi-open, and fully open) to achieve accurate identification of apple flower. Nevertheless, with the continuous advancement of flower thinning devices, the real-time performance of flower identification models is increasingly important.
In recent years, as single-stage object detection algorithms with excellent real-time performance in the field of deep learning, the SSD and YOLO series models have been extensively utilized in the area of flower recognition. For instance, Tian et al. [
21] applied the SSD (Single Shot MultiBox Detector) deep learning technology to the field of flower detection and recognition, and the processing time of a single image by the trained model was 0.13 s. Wu et al. [
22] and Shang et al. [
23] developed real-time apple flower detection models using the YOLOv4 and YOLOv5s deep learning algorithms, respectively. These models provide valuable references for orchard yield estimation and the development of flower thinning machinery. Li et al. [
24] employed YOLOv4 to precisely and rapidly detect kiwifruit flowers and buds simultaneously (with a recognition rate of 38.6 ms per piece and a
mAP of 97.6%), contributing to the development of kiwifruit pollinator robots. Jing et al. [
25] used the improved Sunflower-YOLO to effectively detect the open, half-open, and bud growth stages of sunflowers (model size was 13.8 MB, FPS was 188.52). Bai et al. [
26] proposed a real-time detection model for strawberry flowers and fruits based on YOLOv7 (the frame rate of the improved algorithm was 45 f/s), laying a foundation for the sparse flowers and fruits of strawberries in the greenhouse. Chen et al. [
27] identified apple flowers (buds and flowers) in different growth stages by using an improved YOLOv7 model. The model was capable of accurately and quickly identifying apple flowers in natural scenes (with an accuracy rate of 80.1% and a recognition speed of 42.58 f/s). Although the aforementioned research has classified and identified the different growth stages of the target and can also meet the requirements of real-time detection, the strategy for the subsequent practical operation is not provided.
The formulation of the flower thinning strategy is the continuation of the accurate identification of the model. And it is the prerequisite for realizing the accurate flower thinning operation. For example, Zhou et al. [
28] utilized YOLOv4 to detect pear flowers, optimized the density peak clustering algorithm, and achieved density classification of pear flowers, providing technical support for intelligent pear thinning technology. By enhancing the YOLOv8-Seg model, Si et al. [
29] were capable of offering a reference for the thinning of apple single-branch inflorescences. However, it does not involve coordinate-based spatial thinning decisions. The above research still has certain limitations for achieving accurate flower thinning in natural scenes. Furthermore, most existing research has focused on flower clusters such as apple flowers and pear flowers, with relatively little research on flat peach inflorescence. In fact, a detailed delineation of each growth stage is essential for flat peach inflorescence. This can accurately assess flowering intensity [
20], determine flowering time, and provide scientific guidance for flower thinning operations.
To address the above situation, an object detection model (RBCN-YOLO) specifically designed for flat peach inflorescence was proposed. This model is based on the YOLOv8 framework by optimizing the neck network architecture, integrating an attention mechanism, and enhancing the loss function. Building upon this foundation, a flat peach flower thinning strategy was proposed. This study presents a systematic detection of flat peach inflorescences across multiple phenological stages in complex natural environments, and directly combines the detection results with the flower sparsity strategy. During this process, this study quantitatively characterized the inflorescence at the initial flowering stage. The novelty of this study lies in this comprehensive closed-loop solution, with RBCN-YOLO serving as the core detection component. Specifically, the contributions of this study are primarily reflected in the following aspects:
Data on flat peach inflorescences were collected in a complex natural environment, covering the entire phenological period from the bud stage, the initial flowering stage, to the flowering stage.
Based on the YOLOv8 network, combined with the RepBlock and BiFusion modules, the CAFM module, and the NWD loss function, a detection model capable of accurately identifying peach blossom sequences in different phenological periods was constructed. It has also been deployed and verified on edge devices.
Based on the output information of the detection model, combined with the density classification algorithm, intra-flower membership degree classification, and interval thinning requirements, a new pruning strategy for almond inflorescences was proposed.
By integrating the model recognition effect, heat map, and flower thinning strategy, a visualization system supporting orchard management was designed.
4. Discussion
This research proposes a precise flower thinning strategy for flat peach inflorescence based on RBCN-YOLO, which is deeply focused on achieving precise flower thinning. In the field of precise flower thinning, numerous scholars have contributed a series of inspiring algorithms and ideas, laying a solid foundation for our work and providing valuable references. For instance, Zhou et al. [
28] effectively classified the density of pear flowers by enhancing the density peak clustering algorithm. They utilized a deep learning model to accurately extract pear flower location information and scientifically classify local pear flowers in their natural environment. However, while the classification method is highly effective, it does not address the specific flower thinning strategy post-classification. In contrast, our study builds upon this foundation and clearly outlines a feasible flower thinning protocol. Another noteworthy study is the method proposed by Si et al. [
29] for thinning single-branch inflorescence of apple trees using YOLOv8-Seg. This method addresses the challenge of inflorescence thinning on a single branch of apple trees through semantic segmentation technology. Although this method performs well in specific scenarios, its application is relatively limited, primarily to the processing of single branches. Conversely, our proposed flower thinning strategy based on RBCN-YOLO demonstrates broader applicability. It integrates a target detection deep learning model with a relatively fast detection rate and successfully extends the flower thinning operation to larger scenes, beyond the thinning of a single branch. Importantly, empirical validation has shown that the accuracy of our flower thinning strategy reaches 78.84% of the manual thinning effect, thereby confirming its effectiveness and practicality.
In summary, the precise flower thinning strategy for flat peach inflorescence based on RBCN-YOLO not only offers a novel perspective and approach to precise flower thinning research but also provides a robust reference for achieving intelligent agricultural production.
Limitations
Although the RBCN-YOLO model and flower thinning strategy proposed in this study performed well in the task of almond inflorescence detection, the following limitations were recognized.
The training and validation data of this model were collected from a specific orchard and the flowering period of the same year. Although the dataset contains diverse lighting, angles, and occlusion conditions, its generalization ability still needs further verification under different geographical regions, climatic conditions, or planting patterns. However, enhanced robustness against conditions such as backlighting by improving the model structure and loss function, and simulated illumination variations in data augmentation, extreme or abnormal lighting conditions (such as strong overexposure at noon or extremely low illuminance at dawn and dusk) may still pose challenges to detection accuracy. Although the model has achieved real-time detection on LubanCat-5 and Jetson Nano, its power consumption and thermal management are crucial in long-duration, large-scale field operations. We have not yet conducted a quantitative assessment of the power consumption and heat dissipation of the equipment under continuous full-load operation, which is related to the equipment selection and system durability in practical applications.
Acknowledging these limitations does not diminish the contribution of this study but rather aims to more clearly define its scope of application and provide clear improvement goals for future research. The subsequent research work will focus on the above limitations.
5. Conclusions
In this study, the RBCN-YOLO model was proposed to precisely identify each growth stage of the flat peach inflorescence, particularly the alabastrum stage between bud and flower. This model provided technical support for subsequent flower thinning operations and other orchard management practices. To strengthen the core architecture of the model, the RepBlock and BiFusion modules were incorporated into the model’s neck network, which enhanced the efficiency of the model in extracting context information and fusing features. Meanwhile, by integrating the CAFM attention mechanism into the backbone structure of the model, the accuracy of feature extraction was further improved. Additionally, the NWD measure was combined with the CIoU loss function, which greatly enhanced the detection sensitivity of the model for small-sized targets. Through systematic comparative analysis and ablation experiment verification, the RBCN-YOLO model demonstrated performance improvement in the detection task of flat peach inflorescence. Specifically, the mAP@0.5 value reached 82.9%, the precision (P) was 78.3%, the recall rate (R) was 79.5%, and the F1 score was 78.9%. Especially in the detection of the alabastrum category, the mAP@0.5 was as high as 70.7%. Moreover, it demonstrated good real-time performance on an edge device (LubanCat-5), and the single image detection time is reduced by 12.31% compared with the original model. Furthermore, the flower thinning strategy based on RBCN-YOLO provides a scientific basis for the intelligent operation of the flower thinning robot and achieves an accurate flower thinning rate of 78.84%. And the integrated visualized system combining object detection and flower thinning strategy provides a valuable reference for orchard management.
In future studies, we aim to introduce depth image technology to address the sensitivity of the model to changes in light. This will further enhance the accuracy of flat peach inflorescence detection and reduce reliance on high-quality labelled data. The SAHI was combined with the RBCN-YOLO model to systematically evaluate its recall improvement effect on the bud and snowflake categories on the dataset. It is planned to deploy the RBCN-YOLO model on the orchard inspection robot platform for on-site verification. Furthermore, in future work, we will evaluate the stability and energy efficiency ratio of the model under long-term high-load inference tasks.