Next Article in Journal
Design and Experiment of Comb-Type Header for Plot Breeding Wheat Harvester Based on EDEM
Previous Article in Journal
Impact of Different Groups of Active Substances for Fungal Control on Vineyard Soil Microbiota and Pesticide Residue Profiles
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Instance Segmentation Method for ‘Yuluxiang’ Pear at the Fruit Thinning Stage Based on Improved YOLOv8n-seg Model

1
College of Agricultural Engineering, Shanxi Agricultural University, Jinzhong 030801, China
2
Dryland Farm Machinery Key Technology and Equipment Key Laboratory of Shanxi Province, Jinzhong 030801, China
3
Pomology Institute, Shanxi Agricultural University, Jinzhong 030801, China
*
Author to whom correspondence should be addressed.
Agriculture 2026, 16(3), 346; https://doi.org/10.3390/agriculture16030346
Submission received: 16 November 2025 / Revised: 19 January 2026 / Accepted: 26 January 2026 / Published: 30 January 2026
(This article belongs to the Section Artificial Intelligence and Digital Agriculture)

Abstract

Accurate detection and segmentation of young ‘Yuluxiang’ pear fruits at the fruit thinning stage are crucial for the development of intelligent fruit thinning robots. To address the challenges in recognition and segmentation of young ‘Yuluxiang’ pears in natural environments characterized by occlusion, overlap, and small targets, this paper proposes an improved instance segmentation model based on YOLOv8n-seg, named YOLOv8n-DSW. Firstly, the C2f modules were optimized by introducing DualConv to construct C2f-Dual modules, which enhanced feature extraction capability while reducing the number of parameters. Secondly, a Spatial-Channel Synergistic Attention (SCSA) mechanism was embedded ahead of the small-object detection head to improve detection accuracy for small targets. Finally, the original CIoU loss function was replaced with the WIoU v3 loss function to accelerate model convergence and improve accuracy. Deployment on a Firefly ROC-RK3588S-PC development board confirmed the model’s suitability for edge devices. Experimental results demonstrated that YOLOv8n-DSW achieved excellent performance. The mAP50, mAP75, and mAP50:95 for detection reached 95.6%, 83.2%, and 70.3%, respectively, and those for segmentation were 94.8%, 78.2%, and 65.3%. The proposed model outperformed its baseline, YOLOv8n-seg, as well as other classic models such as YOLOv5n-seg, YOLOv11n-seg, and YOLOv12n-seg. These results demonstrate that YOLOv8n-DSW provides accurate and efficient segmentation of young ‘Yuluxiang’ pear fruits.

1. Introduction

The ‘Yuluxiang’ pear is a high-quality, mid-maturing, and storage-resistant new cultivar of the Korla fragrant pear, characterized by its large fruit size, juicy flesh, and high yield. It was developed through hybridization using the Korla fragrant pear as the female parent and the Snowflake pear as the male parent [1]. ‘Yuluxiang’ pear trees tend to form flower buds easily and exhibit a high fruit set rate. Timely and effective flower and fruit thinning is crucial for ensuring final fruit quality [2,3]. Current primary fruit thinning methods include chemical thinning and manual thinning. Chemical thinning is cost-effective but raises concerns about pesticide residues. While manual thinning allows selective fruit removal by experienced workers to ensure quality, it suffers from low efficiency, high labor demands, and significant time costs [4]. To balance efficiency and cost, developing intelligent fruit thinning equipment has been increasingly adopted in modern orchard management [5], for which accurate identification is a prerequisite.
Currently, existing traditional machine learning methods for fruit recognition include Fuzzy C-Means [6,7], K-means clustering [8], and Support Vector Machines [9,10]. While these methods can achieve fruit recognition, they are susceptible to environmental interference such as varying illumination and leaf occlusion, leading to insufficient recognition accuracy in complex orchard environments. In contrast, deep learning algorithms have demonstrated notable advantages in agricultural detection [11,12,13] by automating feature extraction and offering superior generalization, accuracy, and processing speed. The YOLO series models feature impressive real-time detection performance and high accuracy, which lend them broad application value in agricultural scenarios. For accurate detection of young apple fruits, Song et al. [14] proposed a target detection method based on the YOLOv4-SENL model, which integrated SE Block and NL Block visual attention mechanisms to recalibrate high-level features from both channel and non-local perspectives. The improved model achieved an mAP of 96.9% for young apples in natural environments, demonstrating excellent detection performance. To address the issues of dense grape clusters and occlusion in natural environments, Zhang et al. [15] incorporated CBAM into YOLOv5 to produce YOLOv5-GAP, which achieved an mAP of 95.13%. For multi-task segmentation in pitaya orchards, Zhao et al. [16] enhanced the YOLOP backbone by embedding a Focus and ACmix (FAC) module, integrating C2f and SPPF structures, and optimizing the loss function. Similarly, Li et al. [17] integrated residual modules with gated convolution and a spatial pyramid structure, achieving a segmentation accuracy of 95.11% for green apples. Despite the progress made by existing studies in fruit recognition and segmentation, research, specifically on young fruits, remains relatively scarce, especially when considering the challenges posed by factors like occlusion, overlap, and small targets.
In response to the challenges outlined above, this study introduced the YOLOv8n-DSW model. The main contributions include the following:
(1) This study proposed an instance segmentation model capable of accurately identifying and segmenting young ‘Yuluxiang’ pear fruits under natural conditions. The model incorporates three key improvements: the C2f-Dual module to strengthen feature extraction while achieving model lightweighting; the SCSA mechanism to improve the detection capability for young fruits; and the WIoU v3 loss function to optimize bounding-box regression and accelerate convergence.
(2) Deployment on a Firefly ROC-RK3588S-PC edge computing platform demonstrated that YOLOv8n-DSW maintained high detection and segmentation accuracy while achieving low inference latency on the tested hardware. These results suggest the model’s feasibility for integration into real-time, resource-constrained intelligent thinning equipment and support further development toward automated orchard management.

2. Materials and Methods

2.1. Dataset

2.1.1. Data Acquisition

The images were acquired in early May 2025 at the ‘Yuluxiang’ pear experimental base of the Pomology Institute at Shanxi Agricultural University, Jinzhong City, Shanxi Province. Image acquisition was performed using an Intel® RealSense™ Depth Camera D435i (Intel Corporation, Santa Clara, CA, USA). To ensure data diversity, the images were captured under both frontlighting and backlighting conditions. After removing duplicate and blurred images, a final dataset of 1715 images of young ‘Yuluxiang’ pear fruits was established. The dataset was then randomly split into training, validation, and test sets at a ratio of 7:2:1. The dataset encompassed various scenarios, including frontlighting, backlighting, fruit overlap, and occlusion by branches or leaves, as shown in Figure 1.

2.1.2. Data Processing

To mitigate overfitting caused by a limited number of training images, the training set was augmented, increasing its size from the original 1200 to 6000 images. The augmentation techniques employed included Gaussian noise addition, salt-and-pepper noise addition, brightness adjustment, and horizontal flipping. Examples of the augmented images are shown in Figure 2. The detailed distribution of the dataset is presented in Table 1.
The images of young fruits were manually annotated using the LabelMe software (Version 5.6.1). Each fruit instance was assigned the category label “pear”, and its contour was delineated with polygons, resulting in JSON files containing the respective target information [18,19]. The annotation examples are shown in Figure 3.

2.2. Research Method

2.2.1. YOLOv8 Architecture

YOLOv8 [20], developed by Ultralytics, is a real-time object detection architecture comprising a backbone, a neck, and a detection head. It supports various computer vision tasks, including object detection, instance segmentation, and pose estimation, catering to diverse application scenarios. The backbone, based on CSPDarknet, efficiently extracts multi-scale features from input images, supplying both detailed and high-level semantic information for subsequent processing. The neck employs a Path Aggregation Network with Feature Pyramid Network (PAN-FPN) architecture, which enhances multi-scale feature integration through cross-level connections and effectively balances semantic and spatial details. Furthermore, the architecture employs an anchor-free, decoupled detection head that directly predicts class probabilities and bounding-box coordinates from feature map grids, and a segmentation (mask) head that generates per-instance masks. This design reduces computational complexity while improving localization and segmentation accuracy, enabling efficient real-time instance segmentation. The YOLOv8 framework encompasses multiple versions. Among them, the lightweight YOLOv8n-seg model was chosen as the baseline for this study because it offers a good balance of accuracy and speed and has been widely applied in computer vision.

2.2.2. Improved YOLOv8n-Seg Network Architecture

This study proposed an improved version named YOLOv8n-DSW, based on the YOLOv8n-seg model, and its architecture is depicted in Figure 4. The improvements encompassed three key aspects. First, a DualConv structure was introduced into the original C2f module, forming a C2f-Dual module. This module processed the same input feature map channels simultaneously with dual convolutional kernels, which enhanced spatial feature extraction while reducing both the parameter count and computational cost. Second, to address the challenges of small target size and feature loss in young ‘Yuluxiang’ pears, an SCSA mechanism was embedded before the small-object detection head. This integration enhanced the model’s focus on local details and feature extraction. Third, the WIoU v3 loss function was employed, which strategically focused on anchor boxes of ordinary quality. This replacement accelerated model convergence and enhanced model performance.

2.2.3. C2f-Dual Module

DualConv [21] is a novel convolutional module that processes the same input feature map channels in parallel using two distinct kernel sizes: 3 × 3 and 1 × 1. The 3 × 3 kernels capture spatial features, while the 1 × 1 kernels perform feature fusion and reduce parameters. It also employs group convolution, which partitions the input and output feature maps into multiple independent groups. The convolutional filters within each group process only their corresponding partitioned input channels. This design markedly reduces model complexity and enhances the efficiency of information flow and feature extraction, while maintaining the network’s representational power. The architecture of the DualConv module is depicted in Figure 5.
Assume that the size of the output feature map is Do × Do × N, where Do is the width and height dimension of the output feature map, and the total number of FLOPs performed in a standard convolutional layer FLSC is as follows:
FL SC =   D o 2 × K 2 × M × N
where K × K is the convolutional kernel size, M is the number of input channels, and N is the number of convolutional filters and also the number of output channels.
In DualConv, the number of convolutional filter groups G is used to control the proportion of K × K convolutional kernels in a convolutional filter. For a given G, the proportion of combined simultaneous convolutional kernels with size of (K × K + 1 × 1) is 1/G of all channels, while the proportion of the remaining 1 × 1 convolutional kernels is (1 − 1/G). Therefore, in a dual convolutional layer composed of G convolutional filter groups, the total number of FLOPs is as follows:
FL CC = D o 2 × K 2 × M × N + D o 2 × M × N G
FL PC = D o 2 × M × N   ×   1     1 / G  
FL DC = FL CC + FL PC = D o 2 × K 2 × M × N G + D o 2 × M × N
The computational reduction ratio RDC/SC of the dual convolution layer compared with the standard convolution layer is as follows:
R DC / SC = FL DC FL SC = 1 G + 1 K 2
According to Formula (5), given that K = 3 in the DualConv design, the speedup can reach 8 to 9 times when G is large. Thus, DualConv achieves higher efficiency while maintaining accuracy.
The C2f-Dual module proposed in this study embeds DualConv into the original C2f structure. The architecture of the C2f-Dual module is illustrated in Figure 6, and its workflow is as follows: First, the number of feature map channels is doubled through a 1 × 1 convolution. The input is then split into two parts. One part is processed through several DualBottleneck modules to extract deep features. Subsequently, the output from these modules is concatenated with the other part, which was held as a residual connection. Finally, the concatenated features are passed through a convolutional layer to compress them to the required output channel size.
The DualBottleneck module is illustrated in Figure 7. It first reduces the channel dimension by half using a 3 × 3 convolution and then processes the features through the DualConv layer. This structure enhances the performance of the feature extraction network while maintaining computational efficiency and training stability. Furthermore, by incorporating a cross-stage partial connection mechanism, the C2f-Dual module improves contextual information extraction and synergistically enhances multi-scale feature representation. This design enhances model performance, enabling accurate identification and segmentation of young fruits in natural orchard environments.

2.2.4. SCSA Module

The SCSA module comprises two key components: Shared Multi-Semantic Spatial Attention (SMSA) and Progressive Channel-wise Self-Attention (PCSA) [22]. It integrates the spatial attention’s ability to focus on target locations with the channel attention’s capability to screen critical features, thereby enhancing performance in visual tasks. Specifically, the SMSA extracts multi-semantic spatial information and injects this discriminative spatial information into the PCSA through a progressive compression strategy, which guides the subsequent channel recalibration. The PCSA then utilizes a self-attention mechanism to compute channel-wise similarities, thereby reducing the semantic discrepancies among the different sub-features produced by the SMSA.
As illustrated in Figure 8, the operational workflow is as follows: The SMSA first decomposes the input feature map along the height and width dimensions. After applying global average pooling, it partitions the representation into multiple independent sub-features. These sub-features are subsequently processed by multi-scale MS-DWConv1d, normalized via GroupNorm, and activated by the Sigmoid function to generate the spatial attention map. For the PCSA, it first performs average pooling and GroupNorm normalization on the features processed by the SMSA. It then generates the Query, Key, and Value through a multi-branch depthwise convolution (DWConv). Following feature aggregation via Channel-wise Single-Head Self-Attention (CA-SHSA), the output is passed through an average pooling layer and a sigmoid activation function to produce the channel attention map. Finally, this channel attention map is multiplied by the SMSA-processed features to yield the module’s final output.
The SCSA module is placed before the small-object detection head, as the feature maps for this head have the highest spatial resolution and the smallest receptive fields, making them well-suited to the SCSA mechanism for refining fine-grained features.

2.2.5. WIoU Loss Function

The original YOLOv8n-seg model employs the CIoU loss for bounding-box regression. Although CIoU improves regression accuracy by incorporating aspect ratio, it often leads to unstable gradients and slow convergence when predicted boxes have low overlap with ground truths and large aspect ratio differences. To overcome these limitations, this study adopts WIoU v3 [23]. WIoU v3 incorporates a dynamic non-monotonic focusing mechanism and a rational gradient gain allocation strategy. It assigns smaller gradient gains to anchor boxes with high outlierness, thereby preserving anchor box quality and reducing the impact of harmful gradients.
R WIoU = exp   x x gt   2 +   y y gt   2   W g 2 + H g 2   *
L WIoUv 1 = R WIoU L IoU
β = L IoU *   L IoU ¯ [ 0 , + )
r = β δ α β δ
L WIoUv 3 = r L WIoUv 1
where LIoU is the bounding-box loss, RWIoU is the distance loss, x and y are the coordinates of the predicted bounding-box center, xgt and ygt are the coordinates of the ground truth bounding-box center, Wg and Hg are the width and height of the smallest enclosing rectangle covering both the predicted and ground truth boxes. The operator indicates that Wg and Hg are detached from the computation graph, r represents the gradient gain, β denotes the outlier measure, and α and δ are hyperparameters. L IoU * indicates that LIoU is detached from the computation graph, and L IoU ¯ is the exponential running average.

3. Experiments and Results

3.1. Experimental Setup

3.1.1. Implementation Environment and Training Parameters

All experiments were conducted using PyTorch 2.6.0 with Python 3.9.23, employing the Ultralytics YOLO library version 8.3.91 for model implementation. Development was performed in PyCharm 2024.1 on a Windows 11 workstation equipped with an NVIDIA RTX A6000 GPU (48 GB VRAM) and an Intel Xeon W-2295 CPU @ 3.00 GHz.
The training hyperparameters are summarized in Table 2. The input image size was set to 640 × 640 pixels. The model was trained for 200 epochs with a batch size of 128. The optimizer was Stochastic Gradient Descent (SGD) with an initial learning rate of 0.01 and a weight decay coefficient of 0.0005. The random seed was set to 0. For evaluation and inference, the confidence and Non-Maximum Suppression (NMS) Intersection over Union (IoU) thresholds were set to 0.25 and 0.45, respectively.

3.1.2. Evaluation Metrics

To evaluate the model’s performance on the ‘Yuluxiang’ pear segmentation task, this study employed precision (P), recall (R), mean average precision at an IoU of 0.5 (mAP50), mAP at 0.75 (mAP75), mAP averaged over IoUs from 0.5 to 0.95 (mAP50:95), giga floating-point operations per second (GFLOPs), parameters (Params), inference time, and frames per second (FPS) as evaluation metrics. The formulas for the primary evaluation metrics are defined as follows:
P   =   TP TP + FP × 100 %
R = TP TP + FN × 100 %
mAP 50 = 1 N i = 1 N AP i IoU = 0.5
mAP 75 = 1 N i = 1 N AP i IoU = 0.75
mAP 50 : 95 = 1 N i = 1 N 1 10 k = 0 9 AP i IoU = 0.5 + 0.05 k
where TP, FP, and FN represent the number of true positives, false positives, and false negatives, respectively.

3.2. Experiments

3.2.1. Impact of Integrating the C2f-Dual Module at Different Locations

To validate the effectiveness of the C2f-Dual module, it was used to replace the original C2f modules in both the backbone and neck networks. Three comparative model configurations were established: YOLOv8n-Db (replacement only in the backbone), YOLOv8n-Dn (replacement only in the neck), and YOLOv8n-Dbn (replacement in both). The results are presented in Table 3.
The YOLOv8n-Db model outperformed the baseline model. In terms of detection performance, its mAP50, mAP75, and mAP50:95 were increased by 1.6%, 3.9%, and 2.2%, respectively. For segmentation performance, the corresponding metrics were improved by 1.5%, 2.5%, and 2.3%, respectively. Furthermore, both GFLOPs and Params were reduced. These results indicate that replacing the C2f modules in the backbone effectively enhanced the feature extraction capability, achieved model lightweighting, and improved detection and segmentation precision simultaneously. The YOLOv8n-Dn model, which only modified the C2f modules in the neck, also exhibited better performance than the baseline model in all metrics. The YOLOv8n-Dbn model achieved outstanding performance. Its detection mAP50, mAP75, and mAP50:95 reached 94.8%, 82.5%, and 69.7%, while its segmentation mAP50, mAP75, and mAP50:95 reached 94.0%, 75.3%, and 64.5%, all of which were higher than those of the other two models and the baseline model. Meanwhile, its GFLOPs and Params stood at 11.3 and 2.96 M, corresponding to the lowest computational complexity and parameter count among all models. In summary, the comprehensive replacement of C2f modules in both the backbone and the neck achieved the optimal balance between accuracy improvement and model lightweighting.

3.2.2. Analysis of SCSA Module Integration Effects at Different Detection Layers

To analyze the impact of integrating the SCSA module at different positions within the detection head, the SCSA module was individually added to the large, medium, and small-object detection layers of the YOLOv8n-seg model. The resulting models are denoted as YOLOv8n-L, YOLOv8n-M, and YOLOv8n-S, respectively. A model incorporating the SCSA module across all three detection layers is denoted as YOLOv8n-all. Table 4 presents the performance metrics of these five models.
As shown in Table 4, integrating the SCSA module individually into the large, medium, or small-object detection layers improved most key metrics for both detection and segmentation tasks compared with the original YOLOv8n-seg model. The most noticeable enhancements were observed when the module was added to the small-object detection layer. For the detection task, precision, recall, mAP50, mAP75, and mAP50:95 increased by 0.7%, 1.7%, 1.7%, 3.1%, and 1.9%, respectively. In the segmentation task, the corresponding metrics improved by 1.2%, 2.0%, 2.2%, 3.5%, and 2.6%, indicating a more pronounced benefit for segmentation performance. This can be attributed to the small-object detection layer processing shallow features from the backbone network, characterized by larger feature map resolutions and smaller receptive fields. Furthermore, the overall performance of the YOLOv8n-all model was slightly inferior to that of YOLOv8n-S. Consequently, the final model configuration integrates the SCSA module solely into the small-object detection layer.

3.2.3. Comparative Experiments with Different Loss Functions

To validate the performance of the WIoU loss function in the detection and segmentation tasks of young ‘Yuluxiang’ pear fruits, both CIoU and WIoU were, respectively, used as the loss functions for training the YOLOv8n-seg model. Figure 9 illustrates the decrease in loss values for both loss functions on the validation set.
Both box loss curves drop rapidly in the initial training phase, indicating quick weight adaptation to reduce bounding-box errors. The CIoU loss is higher than the WIoU loss in the early epochs, showing that WIoU optimizes bounding-box regression more quickly; over the full run, WIoU remains lower and stabilizes, suggesting better validation performance for bounding-box regression. The mask loss curves for both losses decline and converge to nearly identical values, implying that WIoU’s main benefit is improved bounding-box regression rather than a direct effect on segmentation; more accurate candidate regions produced by WIoU indirectly favor segmentation and yield observable metric gains.
WIoU also exhibits faster convergence than CIoU in the box loss curves, enabling quicker parameter optimization and lower loss values. These properties suggest that WIoU can more effectively guide training, enhancing stability and reducing overfitting. In summary, WIoU accelerates convergence and improves regression accuracy, leading to better overall model performance.

3.2.4. Ablation Experiment and Result Analysis

Ablation experiments were conducted to assess the impact of the proposed improvements. The results are summarized in Table 5.
Incorporating the DualConv architecture into the baseline YOLOv8n-seg improved both detection and segmentation while reducing model complexity. Detection mAP50, mAP75, and mAP50:95 reached 94.8%, 82.5%, and 69.7%, corresponding to gains of 1.8%, 4.0%, and 2.3% over the baseline; segmentation mAP50, mAP75, and mAP50:95 increased by 1.9%, 2.8%, and 2.4%, respectively. Params and GFLOPs also decreased, indicating that DualConv both strengthens feature extraction and achieves lightweighting by using dual convolutional kernels. Adding the SCSA module before the small-object detection layer improved detection mAP50 by 1.7% and segmentation mAP50 by 2.2% relative to the baseline. The adoption of the WIoU enhanced model performance. For the detection task, mAP50, mAP75, and mAP50:95 reached 95.3%, 82.2%, and 69.7%, respectively. Segmentation metrics also showed substantial improvement, while Params and GFLOPs remained unchanged.
When any two improvements were combined, the model’s overall performance remained better than the baseline. The simultaneous integration of all three enhancements yielded optimal performance. The final model achieved detection precision, recall, mAP50, mAP75, and mAP50:95 of 94.7%, 90.2%, 95.6%, 83.2%, and 70.3%, respectively, representing gains of 1.3%, 2.5%, 2.6%, 4.7%, and 2.9% over the baseline. For segmentation, it attained precision, recall, mAP50, mAP75, and mAP50:95 of 95.0%, 89.9%, 94.8%, 78.2%, and 65.3%, improving by 2.5%, 3.4%, 2.7%, 5.7%, and 3.2%, respectively. These results validate the synergistic value of the three strategies: the lightweight feature extraction of DualConv, the small-target focus of SCSA, and the loss optimization of WIoU collectively enhance the model’s adaptability to growth scenarios of young ‘Yuluxiang’ pear fruits.
Figure 10 compares the segmentation results of the baseline model and the proposed YOLOv8n-DSW model to illustrate the improvement in performance. The baseline model exhibits several limitations: a high rate of missed detections, particularly for small or heavily occluded young fruits (Figure 10b–d); false positives caused by leaves being misclassified as young fruits (Figure 10c,d); and insufficient segmentation accuracy resulting in coarse delineation of fruit boundaries (Figure 10d). In contrast, the proposed YOLOv8n-DSW model, incorporating all three improvements, accurately segments occluded fruits and more precisely localizes small targets. It better captures fine-grained features and improves boundary segmentation, producing clear and well-defined contours. Overall, the proposed model demonstrates improved robustness and generalization in handling occluded and overlapping targets.
The experimental results confirm that the proposed method effectively improves model performance, yielding favorable results in the detection and segmentation of young ‘Yuluxiang’ pear fruits.

3.2.5. Heatmap Analysis

To better visualize the enhancement in the model’s recognition capability, this study utilizes Gradient-weighted Class Activation Mapping (Grad-CAM) to generate heatmaps [24], enabling detailed examination of the network’s attention distribution. Grad-CAM computes the gradients of the target class score relative to the feature maps in the final convolutional layer. These gradients undergo global average pooling to obtain weights for each feature map. The weights are then used for a weighted summation with their corresponding feature maps, followed by ReLU activation. Finally, the result is upsampled to the original image dimensions, producing a heatmap that highlights regions most critical for the model’s predictions.
Figure 11 compares heatmaps generated by the baseline and improved models for young ‘Yuluxiang’ pear images. In Figure 11a,b, the enhanced YOLOv8n-DSW model shows more intense activation in the actual fruit regions, indicating better focus on the targets themselves and demonstrating improved localization precision. In Figure 11c, while the baseline YOLOv8n-seg model fails to activate significantly in the occluded fruit area, the improved model produces distinct heatmap values on the exposed portions of occluded fruits. Additionally, in Figure 11d,e, the baseline model exhibits noticeable activation in background regions without fruits, whereas such background activation is clearly reduced in the improved model. These observations demonstrate that the enhanced YOLOv8n-DSW model possesses superior capability in distinguishing targets from background and identifying occluded fruits, thereby reducing false positives and missed detections. The proposed model exhibits stronger robustness and stability in complex scenarios.

3.2.6. Comparative Analysis of Various Models

To further evaluate the effectiveness of the improved model, YOLOv5n-seg [25], YOLOv9c-seg [26], YOLO11n-seg [27], YOLO12n-seg [28], and the original and improved YOLOv8n-seg models were compared under identical experimental conditions. The comparative results of these six models are summarized in Table 6.
As shown in Table 6, the proposed YOLOv8n-DSW model demonstrates superior overall performance. For object detection, it attains a precision of 94.7%, a recall of 90.2%, a mAP50 of 95.6%, a mAP75 of 83.2%, and a mAP50:95 of 70.3%. These represent clear improvements over its baseline. For instance, in segmentation, it achieves a precision of 95.0%, a recall of 89.9%, a mAP50 of 94.8%, a mAP75 of 78.2%, and a mAP50:95 of 65.3%, also showing consistent gains across all segmentation metrics compared with the baseline.
Notably, compared to newer models like YOLOv11n-seg and YOLOv12n-seg, it delivers the best performance across nearly all key metrics for both detection and segmentation, with only a slight increase in computational cost. Among all the compared models, YOLOv8n-DSW achieves the best balance between performance and efficiency. In terms of accuracy, it is only slightly behind the highest-performing model, YOLOv9c-seg. However, YOLOv9c-seg requires substantial computational resources of 157.6 GFLOPs and 27.6 million parameters. Such a massive computational overhead makes it difficult to deploy in practical scenarios. In contrast, YOLOv8n-DSW maintains this high performance with a computational cost of only 11.3 GFLOPs and 2.97 million parameters.
This confirms that the proposed improvements effectively enhance the model’s capability without introducing extra efficiency overhead. Therefore, YOLOv8n-DSW represents an excellent choice, effectively balancing high precision with practical deployment efficiency.

3.2.7. Visual System Design and Edge Deployment

This study developed a visualization system and verified model deployment on an edge computing platform. The visualization system features a split-panel interface: the left panel shows original input images while the right panel displays model-generated segmentation results. Young fruit targets are annotated with bounding boxes, category labels, and confidence scores. The system supports multiple operational modes, including model weight selection, static image detection, and real-time camera detection, enabling flexible switching between offline analysis and online detection to accommodate various application scenarios. The visual system design and deployment architecture are illustrated in Figure 12.
The implementation adopts the Firefly ROC-RK3588S-PC development board equipped with the Ubuntu 20.04 operating system. To enable the deep learning model to be compatible with the RK3588 AI chip and achieve efficient accelerated neural network inference using the NPU, this study completed the conversion of the GPU-trained model to an NPU-compatible format through model conversion, quantization, and other techniques. Edge-side inference is implemented by calling the underlying RKNN inference engine via RKNN-Toolkit-Lite. The quantization process was as follows: first, the trained best.pt file was converted to the ONNX format; subsequently, the weights and activation values of the model were quantized from 32-bit floating-point (FP32) to 8-bit integer (INT8) using the RKNN Toolkit with a batch size of 1; finally, the RKNN model suitable for NPU-accelerated inference was obtained.
The test results are shown in Table 7. The improved YOLOv8n-DSW model reduces the inference time by approximately 5.7% compared to the original model. As seen in the segmentation results in Figure 13, the improved model produces fewer missed and false detections than the baseline. Deployment on edge devices verifies the practicality of YOLOv8n-DSW in practical agricultural environments. This work provides crucial technical support for the development of fruit thinning robots.

4. Discussion

Experimental results show that the proposed model notably outperforms the baseline YOLOv8n-seg model and other lightweight segmentation models. This advantage is mainly attributed to the synergistic effect of the three improved components. By integrating the C2f-Dual module, the SCSA attention mechanism, and the WIoU v3 loss function, the improved model significantly enhances the accuracy of instance segmentation for young ‘Yuluxiang’ pear fruits in natural orchard environments. Concurrently, the model was deployed on the Firefly ROC-RK3588S-PC edge device, validating its feasibility for real-time perception in resource-constrained environments.
Despite advances in the detection and segmentation of young fruits, occlusion remains a persistent challenge. Severe occlusion of young ‘Yuluxiang’ pears still presents considerable difficulty for detection, often leaving them largely or entirely invisible in captured images.
Future research will focus on three directions. First, the model architecture will be further refined to additionally reduce complexity while maintaining high accuracy. Second, the proposed model will be deployed on an intelligent fruit thinning robotic system to adapt to actual production environments. Finally, the framework will be extended to other small-target products, such as apples during the thinning period, to assess its generalizability and promote wider application in agricultural automation.

5. Conclusions

To address the challenges of occlusion, overlap, and small targets in detecting and segmenting young ‘Yuluxiang’ pears under natural orchard conditions, this study proposed YOLOv8n-DSW, an improved instance segmentation model. The model integrated a lightweight C2f-Dual module for enhanced feature extraction, incorporated an SCSA attention mechanism to improve the recognition capability of small-sized fruits, and used the WIoU v3 loss to optimize bounding-box regression. These enhancements effectively improved segmentation accuracy and addressed challenges posed by occlusion and overlap. Comprehensive experiments showed that the proposed YOLOv8n-DSW outperformed the baseline YOLOv8n-seg and other lightweight models, including YOLOv5n-seg, YOLOv11n-seg, and YOLOv12n-seg, across key performance metrics. Deployment on an edge computing platform validated the model’s efficiency and practical suitability for field applications. This work offers a reliable technical solution for intelligent fruit thinning robots and contributes to the advancement of automated, precision orchard management.

Author Contributions

Conceptualization, W.H. and Z.Z.; methodology, W.H., Z.Z., X.Z., and Y.Z.; software, W.H., Y.S., and L.C.; data curation, W.H. and H.L.; investigation, W.H., B.T., and S.Y.; writing—original draft preparation, W.H. and Z.Z.; writing—review and editing, W.H., X.Z., and Z.Z.; funding acquisition, Z.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Scientific and Technological Innovation Programs of Higher Education Institutions in Shanxi, China, grant number 2022L093, the Key Research and Development Program of Shanxi Province, China, grant number 202102020101012, the Corps Science and Technology Planning Project, China, grant number 2024AB040, and the Quwo County National Modern Agricultural Industrial Park Project, China, grant number SXRTFWZB2506-26-3.

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author.

Acknowledgments

The authors would like to extend their thanks to the technical editor and anonymous reviewers for their constructive comments and suggestions on this study.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Xue, L. Intensive Planting and High Yield Cultivation Techniques for Yuluxiang Pear in Hilly and Mountainous Regions. Plant Dis. Pests 2025, 16, 31–33. [Google Scholar]
  2. Zhang, E. Planting and Management Technology of Yulu Fragrant Pear in Yuanqu County. Agric. Technol. Equip. 2023, 188–190. (In Chinese) [Google Scholar]
  3. Ma, X.; Su, X.; Qiao, J.; Zhang, X.; Liu, X. Green and High-Efficiency Cultivation Techniques of ‘Yuluxiang’ Pear in Shenmu, Shaanxi Province. North. Fruits 2025, 28–31. (In Chinese) [Google Scholar]
  4. Long, Y.; Yang, Z.; He, M. Recognizing apple targets before thinning using improved YOLOv7. Trans. Chin. Soc. Agric. Eng. 2023, 39, 191–199. (In Chinese) [Google Scholar]
  5. Luo, X.; Liao, J.; Hu, L.; Zhou, Z.; Zhang, Z.; Zang, Y.; Wang, P.; He, J. Research progress of intelligent agricultural machinery and practice of unmanned farm in China. J. South China Agric. Univ. 2021, 42, 8–17. (In Chinese) [Google Scholar]
  6. Henila, M.; Chithra, P. Segmentation using fuzzy cluster-based thresholding method for apple fruit sorting. IET Image Process. 2020, 14, 4178–4187. [Google Scholar] [CrossRef]
  7. Marlinda, L.; Fatchan, M.; Widiyawati, W.; Aziz, F.; Indrarti, W. Segmentation of Mango Fruit Image Using Fuzzy C-Means. Sink. J. Dan Penelit. Tek. Inform. 2021, 5, 275–281. [Google Scholar] [CrossRef]
  8. Sreekanth, G.; Thangaraj, P.; Kirubakaran, S. Fruit detection using improved K-means algorithm. J. Crit. Rev. 2020, 7, 5–6. [Google Scholar] [CrossRef]
  9. Tang, Y.; Zhou, H.; Wang, H.; Zhang, Y. Fruit detection and positioning technology for a Camellia oleifera C. Abel orchard based on improved YOLOv4-tiny model and binocular stereo vision. Expert Syst. Appl. 2023, 211, 118573. [Google Scholar] [CrossRef]
  10. Venkateswarlu, A.; Veeralakshmi, P. Improved accuracy for finding fruit quality through ripening using convolution neural network, artificial neural networks in comparison with support vector machine. AIP Conf. Proc. 2025, 3267, 020071. [Google Scholar]
  11. Chen, J.; Ji, C.; Zhang, J.; Feng, Q.; Li, Y.; Ma, B. A method for multi-target segmentation of bud-stage apple trees based on improved YOLOv8. Comput. Electron. Agric. 2024, 220, 108876. [Google Scholar] [CrossRef]
  12. Naranjo-Torres, J.; Mora, M.; Hernández-García, R.; Barrientos, R.J.; Fredes, C.; Valenzuela, A. A review of convolutional neural network applied to fruit image processing. Appl. Sci. 2020, 10, 3443. [Google Scholar] [CrossRef]
  13. Varjão, J.O.R.; Botelho, G.M.; Almeida, T.; Martins, G.; Silva, W. Citrus fruit quality classification using support vector machines. Int. J. Advanc. Eng. Res. Sci. 2019, 6, 59–65. [Google Scholar] [CrossRef]
  14. Song, H.; Jiang, M.; Wang, Y.; Song, L. Efficient detection method for young apples based on the fusion of convolutional neural network and visual attention mechanism. Trans. Chin. Soc. Agric. Eng. 2021, 37, 297–303. (In Chinese) [Google Scholar]
  15. Zhang, T.; Wu, F.; Wang, M.; Chen, Z.; Li, L.; Zou, X. Grape-Bunch Identification and Location of Picking Points on Occluded Fruit Axis Based on YOLOv5-GAP. Horticulturae 2023, 9, 498. [Google Scholar] [CrossRef]
  16. Wenfeng, Z.; Yuanjue, H.; Minyue, Z.; Zhenyuan, L.; Zitao, L.; Jiajun, H. Multi-task Visual Perception Method in Dragon Orchards Based on OrchardYOLOP. Nongye Jixie Xuebao/Trans. Chin. Soc. Agric. Mach. 2024, 55, 160–170. [Google Scholar]
  17. Li, Q.; Jia, W.; Sun, M.; Hou, S.; Zheng, Y. A novel green apple segmentation algorithm based on ensemble U-Net under complex orchard environment. Comput. Electron. Agric. 2021, 180, 105900. [Google Scholar] [CrossRef]
  18. Shi, J.; Li, H.; Jing, L.; Xiao, Z.; Zhou, H.; Zhang, Z.; Xu, S. ITC-YOLO: An improved YOLOv8 model for individual tree crown segmentation in complex forest environments. Ecol. Indic. 2025, 180, 114355. [Google Scholar] [CrossRef]
  19. Zhou, J.; Chen, M.; Zhang, M.; Zhang, Z.; Zhang, Y.; Wang, M. Improved YOLOv8 for multi-colored apple fruit instance segmentation and 3D localization. Artif. Intell. Agric. 2026, 16, 381–396. [Google Scholar] [CrossRef]
  20. Varghese, R.; Sambath, M. Yolov8: A novel object detection algorithm with enhanced performance and robustness. In Proceedings of the 2024 International Conference on Advances in Data Engineering and Intelligent Computing Systems (ADICS); IEEE: Piscataway, NJ, USA, 2024; pp. 1–6. [Google Scholar]
  21. Zhong, J.; Chen, J.; Mian, A. DualConv: Dual convolutional kernels for lightweight deep neural networks. IEEE Trans. Neural Netw. Learn. Syst. 2022, 34, 9528–9535. [Google Scholar] [CrossRef] [PubMed]
  22. Si, Y.; Xu, H.; Zhu, X.; Zhang, W.; Dong, Y.; Chen, Y.; Li, H. SCSA: Exploring the synergistic effects between spatial and channel attention. arXiv 2024, arXiv:2407.05128. [Google Scholar] [CrossRef]
  23. Tong, Z.; Chen, Y.; Xu, Z.; Yu, R. Wise-IoU: Bounding box regression loss with dynamic focusing mechanism. arXiv 2023, arXiv:2301.10051. [Google Scholar]
  24. Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-cam: Visual explanations from deep networks via gradient-based localization. In Proceedings of the Proceedings of the IEEE International Conference on Computer Vision; IEEE: Piscataway, NJ, USA, 2017; pp. 618–626. [Google Scholar]
  25. Liang, X.; Wei, Z.; Chen, K. A method for segmentation and localization of tomato lateral pruning points in complex environments based on improved YOLOV5. Comput. Electron. Agric. 2025, 229, 109731. [Google Scholar] [CrossRef]
  26. Zhu, G.; Luo, Z.; Ye, M.; Xie, Z.; Luo, X.; Hu, H.; Wang, Y.; Ke, Z.; Jiang, J.; Wang, W. Instance Segmentation of Sugar Apple (Annona squamosa) in Natural Orchard Scenes Using an Improved YOLOv9-seg Model. Agriculture 2025, 15, 1278. [Google Scholar] [CrossRef]
  27. Xue, S.; Li, Z.; Wang, D.; Zhu, T.; Zhang, B.; Ni, C. YOLO-ALDS: An instance segmentation framework for tomato defect segmentation and grading based on active learning and improved YOLO11. Comput. Electron. Agric. 2025, 238, 110820. [Google Scholar] [CrossRef]
  28. Alzahrani, A.S.; Iqbal, A.; Zafar, W.; Husnain, G. Swin-YOLO-SAM: A hybrid Transformer-based framework integrating Swin Transformer, YOLOv12, and SAM-2.1 for automated identification and segmentation of date palm leaf diseases. Front. Plant Sci. 2025, 16, 1666374. [Google Scholar] [CrossRef]
Figure 1. Images of young ‘Yuluxiang’ pear: (a) frontlight; (b) backlight; (c) fruit overlap; (d) branch occlusion; (e) leaf occlusion.
Figure 1. Images of young ‘Yuluxiang’ pear: (a) frontlight; (b) backlight; (c) fruit overlap; (d) branch occlusion; (e) leaf occlusion.
Agriculture 16 00346 g001
Figure 2. Data augmentation process: (a) original image; (b) Gaussian noise; (c) salt-and-pepper noise; (d) brightness adjustment; (e) horizontal flip.
Figure 2. Data augmentation process: (a) original image; (b) Gaussian noise; (c) salt-and-pepper noise; (d) brightness adjustment; (e) horizontal flip.
Agriculture 16 00346 g002
Figure 3. Examples of dataset annotation details.
Figure 3. Examples of dataset annotation details.
Agriculture 16 00346 g003
Figure 4. Architecture of the proposed YOLOv8n-DSW model. Modifications relative to the original YOLOv8n-seg are indicated below: all C2f modules in the backbone and neck are replaced with the proposed C2f-Dual modules (green blocks); an SCSA attention module is inserted before the small-object detection head (outlined by a red dashed box); and the bounding-box regression loss is replaced with WIoU v3.
Figure 4. Architecture of the proposed YOLOv8n-DSW model. Modifications relative to the original YOLOv8n-seg are indicated below: all C2f modules in the backbone and neck are replaced with the proposed C2f-Dual modules (green blocks); an SCSA attention module is inserted before the small-object detection head (outlined by a red dashed box); and the bounding-box regression loss is replaced with WIoU v3.
Agriculture 16 00346 g004
Figure 5. DualConv network structure.
Figure 5. DualConv network structure.
Agriculture 16 00346 g005
Figure 6. C2f-Dual structure diagram.
Figure 6. C2f-Dual structure diagram.
Agriculture 16 00346 g006
Figure 7. DualBottleneck structure diagram.
Figure 7. DualBottleneck structure diagram.
Agriculture 16 00346 g007
Figure 8. SCSA structure diagram.
Figure 8. SCSA structure diagram.
Agriculture 16 00346 g008
Figure 9. Loss curves of different loss functions: (a) box loss; (b) mask loss.
Figure 9. Loss curves of different loss functions: (a) box loss; (b) mask loss.
Agriculture 16 00346 g009
Figure 10. Examples of model segmentation results: (a) frontlight; (b) fruit overlap; (c) branch occlusion; (d) leaf occlusion. Note: Red circles indicate missed detections of young fruits, yellow circles indicate false positive detections, and green circles highlight instances where the model exhibits inaccurate segmentation.
Figure 10. Examples of model segmentation results: (a) frontlight; (b) fruit overlap; (c) branch occlusion; (d) leaf occlusion. Note: Red circles indicate missed detections of young fruits, yellow circles indicate false positive detections, and green circles highlight instances where the model exhibits inaccurate segmentation.
Agriculture 16 00346 g010
Figure 11. The heatmap comparison of the model before and after improvement: (a) frontlight; (b) backlight; (c) fruit overlap; (d) branch occlusion; (e) leaf occlusion.
Figure 11. The heatmap comparison of the model before and after improvement: (a) frontlight; (b) backlight; (c) fruit overlap; (d) branch occlusion; (e) leaf occlusion.
Agriculture 16 00346 g011
Figure 12. Model deployment on ROC-RK3588S-PC.
Figure 12. Model deployment on ROC-RK3588S-PC.
Agriculture 16 00346 g012
Figure 13. Segmentation results on RK3588S: (a) YOLOv8n-seg; (b) YOLOv8n-DSW.
Figure 13. Segmentation results on RK3588S: (a) YOLOv8n-seg; (b) YOLOv8n-DSW.
Agriculture 16 00346 g013
Table 1. Dataset of young ‘Yuluxiang’ pear.
Table 1. Dataset of young ‘Yuluxiang’ pear.
Data SetTraining SetValidation SetTest SetTotal Quantity
Original image12003431721715
Augmented image4800004800
Total quantity60003431726515
Table 2. Training parameters.
Table 2. Training parameters.
Training ParameterValue
Input Image Size640 × 640
Epochs200
Batch Size128
OptimizerSGD
Initial Learning Rate0.01
Weight Decay Coefficient0.0005
AugmentTrue
Hue Shift (hsv_h)0.015
Saturation Shift (hsv_s)0.7
Brightness Shift (hsv_v)0.4
Translation (translate)0.1
Scale (scale)0.5
Horizontal Flip (fliplr)0.5
Mosaic (mosaic)1.0
Seed0
Table 3. Results of applying C2f-Dual to different positions.
Table 3. Results of applying C2f-Dual to different positions.
ModelsDetectionSegmentationGFLOPsParams
P
/%
R
/%
mAP50
/%
mAP75
/%
mAP50:95
/%
P
/%
R
/%
mAP50
/%
mAP75
/%
mAP50:95
/%
YOLOv8n-seg93.487.793.078.567.492.586.592.172.562.112.03.26 × 106
YOLOv8n-Db93.988.994.682.469.693.188.293.675.064.411.63.10 × 106
YOLOv8n-Dn94.188.794.781.269.193.488.793.975.164.111.73.11 × 106
YOLOv8n-Dbn94.189.194.882.569.793.788.694.075.364.511.32.96 × 106
Table 4. Performance indicators of five models.
Table 4. Performance indicators of five models.
ModelsDetectionSegmentation
P
/%
R
/%
mAP50
/%
mAP75
/%
mAP50:95
/%
P
/%
R
/%
mAP50
/%
mAP75
/%
mAP50:95
/%
YOLOv8n-seg93.487.793.078.567.492.586.592.172.562.1
YOLOv8n-L92.589.294.681.269.293.787.594.275.564.5
YOLOv8n-M93.589.095.081.369.193.587.994.375.164.3
YOLOv8n-S94.189.494.781.669.393.788.594.376.064.7
YOLOv8n-all93.988.994.781.569.093.988.194.275.664.7
Table 5. Ablation experiments’ results.
Table 5. Ablation experiments’ results.
ModelsABCDetectionSegmentationGFLOPsParams
P
/%
R
/%
mAP50
/%
mAP75
/%
mAP50:95
/%
P
/%
R
/%
mAP50
/%
mAP75
/%
mAP50:95
/%
1×××93.487.793.078.567.492.586.592.172.562.112.03.26 × 106
2××94.189.194.882.569.793.788.694.075.364.511.32.96 × 106
3××94.189.494.781.669.393.788.594.376.064.712.03.27 × 106
4××94.689.295.382.269.793.089.194.677.165.012.03.26 × 106
5×93.988.794.979.769.594.488.993.874.564.311.32.97 × 106
6×94.488.994.680.967.994.688.594.475.664.211.73.11 × 106
7×93.889.994.879.368.393.389.994.276.964.912.03.27 × 106
894.790.295.683.270.395.089.994.878.265.311.32.97 × 106
Note: A, B, C denote C2f-Dual, SCSA, WIoU. “√” indicates that the module is included, “×” indicates that the module is not included.
Table 6. Comparison results of different network models.
Table 6. Comparison results of different network models.
ModelsDetectionSegmentationGFLOPsParams
P
/%
R
/%
mAP50
/%
mAP75
/%
mAP50:95
/%
P
/%
R
/%
mAP50
/%
mAP75
/%
mAP50:95
/%
YOLOv5n-seg90.978.585.972.162.490.877.684.368.557.56.71.88 × 106
YOLOv8n-seg93.487.793.078.567.492.586.592.172.562.112.03.26 × 106
YOLOv9c-seg95.490.396.190.879.295.590.596.285.570.9157.62.76 × 107
YOLO11n-seg92.590.494.482.669.393.489.194.477.864.710.42.84 × 106
YOLO12n-seg93.690.795.283.169.993.389.994.277.664.510.42.82 × 106
YOLOv8n-DSW94.790.295.683.270.395.089.994.878.265.311.32.97 × 106
Table 7. Performance comparison of models on the edge device.
Table 7. Performance comparison of models on the edge device.
ModelsPreprocess Time/msInference Time/msPostprocess Time/msFPSmAP50
(box)/%
mAP50:95
(box)/%
mAP50
(Mask)/%
mAP50:95
(Mask)/%
YOLOv8n-seg14.3126.7028.0314.4891.065.990.160.7
YOLOv8n-DSW13.7525.1727.1215.1493.568.892.763.9
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Hao, W.; Zhang, X.; Liang, H.; Shi, Y.; Chen, L.; Tang, B.; Yang, S.; Zhang, Y.; Zhang, Z. Instance Segmentation Method for ‘Yuluxiang’ Pear at the Fruit Thinning Stage Based on Improved YOLOv8n-seg Model. Agriculture 2026, 16, 346. https://doi.org/10.3390/agriculture16030346

AMA Style

Hao W, Zhang X, Liang H, Shi Y, Chen L, Tang B, Yang S, Zhang Y, Zhang Z. Instance Segmentation Method for ‘Yuluxiang’ Pear at the Fruit Thinning Stage Based on Improved YOLOv8n-seg Model. Agriculture. 2026; 16(3):346. https://doi.org/10.3390/agriculture16030346

Chicago/Turabian Style

Hao, Weihao, Xi Zhang, Hao Liang, Yaozong Shi, Lihang Chen, Bo Tang, Sheng Yang, Yanqing Zhang, and Zhiyong Zhang. 2026. "Instance Segmentation Method for ‘Yuluxiang’ Pear at the Fruit Thinning Stage Based on Improved YOLOv8n-seg Model" Agriculture 16, no. 3: 346. https://doi.org/10.3390/agriculture16030346

APA Style

Hao, W., Zhang, X., Liang, H., Shi, Y., Chen, L., Tang, B., Yang, S., Zhang, Y., & Zhang, Z. (2026). Instance Segmentation Method for ‘Yuluxiang’ Pear at the Fruit Thinning Stage Based on Improved YOLOv8n-seg Model. Agriculture, 16(3), 346. https://doi.org/10.3390/agriculture16030346

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop