You are currently viewing a new version of our website. To view the old version click .
Sensors
  • Article
  • Open Access

2 December 2025

Enhanced Image Annotation in Wild Blueberry (Vaccinium angustifolium Ait.) Fields Using Sequential Zero-Shot Detection and Segmentation Models

,
,
,
and
1
Department of Engineering, Faculty of Agriculture, Dalhousie University, Truro, NS B2N 5E3, Canada
2
Atlantic Institute for Digital Agriculture, Dalhousie University, Truro, NS B2N 5E3, Canada
*
Author to whom correspondence should be addressed.
This article belongs to the Special Issue Sensors and Machine Vision Technologies for Stress and Disease Detection in Digital Agriculture

Abstract

This research addresses the critical need for efficient image annotation in precision agriculture, using the wild blueberry (Vaccinium angustifolium Ait.) cropping system as a representative application to enable data-driven crop management. Tasks such as automated berry ripeness detection, plant disease identification, plant growth stage monitoring, and weed detection rely on extensive annotated datasets. However, manual annotation is labor-intensive, time-consuming, and impractical for large-scale agricultural systems. To address this challenge, this study evaluates an automated annotation pipeline that integrates zero-shot detection models from two frameworks (Grounding DINO and YOLO-World) with the Segment Anything Model version 2 (SAM2). The models were tested on detecting and segmenting ripe wild blueberries, developmental wild blueberry buds, hair fescue (Festuca filiformis Pourr.), and red leaf disease (Exobasidium vaccinii). Grounding DINO consistently outperformed YOLO-World, with its Swin-T achieving mean Intersection over Union (mIoU) scores of 0.694 ± 0.175 for fescue grass and 0.905 ± 0.114 for red leaf disease when paired with SAM2-Large. For ripe wild blueberry detection, Swin-B with SAM2-Small achieved the highest performance (mIoU of 0.738 ± 0.189). Whereas for wild blueberry buds, Swin-B with SAM2-Large yielded the highest performance (0.751 ± 0.154). Processing times were also evaluated, with SAM2-Tiny, Small, and Base demonstrating the shortest durations when paired with Swin-T (0.30–0.33 s) and Swin-B (0.35–0.38 s). SAM2-Large, despite higher segmentation accuracy, had significantly longer processing times (significance level α = 0.05), making it less practical for real-time applications. This research offers a scalable solution for rapid, accurate annotation of agricultural images, improving targeted crop management. Future research should optimize these models for different cropping systems, such as orchard-based agriculture, row crops, and greenhouse farming, and expand their application to diverse crops to validate their generalizability.

1. Introduction

Image annotation is pivotal in various agricultural practices, including creating prescription maps, identifying plant species and diseases, and analyzing growth stages [1,2,3,4]. With advancements in machine learning and computer vision, the demand for rapid and accurate data annotation has increased. Training machine learning models often requires thousands of precisely annotated images, a task that can be time-consuming [5,6]. In wild blueberry (Vaccinium angustifolium Ait.) cultivation, image annotation aids in detecting ripe berries, monitoring growth stages, and identifying disease infections and weeds. Accurate fruit identification supports yield estimation, quality assessment, and ripeness determination [7,8]. However, this task requires experienced personnel trained in feature recognition. The ability to identify weeds and diseases allows management practices such as variable rate application of agrochemicals to be used [9,10]. Applied at various stages during wild blueberry growth, agrochemicals such as herbicides and fungicides are essential for weed and disease prevention [11,12]. Identifying the bud stage, crucial for optimal fungicide application, can be challenging for inexperienced personnel but is essential for maximizing yield.
Wild blueberries are grown in a 2-year cropping cycle with the first year being the sprout year and the second year being the harvest year [13]. Throughout these two years, fungicides, herbicides and insecticides are applied at specific times within the cropping cycle based on the growth stage of the wild blueberry plants [14,15]. Growth stages along with other factors such as growing degree days are also used for determining the correct timings for pollination and harvest [15]. The percentage of buds in F2 stage of development can help producers ensure proper timing when managing their fields [16]. Approximately 5% of a wild blueberry field is in the F2 stage at 110 growing degree days while 80% of the field is in the F2 stage at 222 growing degree days [16]. Wild blueberry fields are harvested at approximately 1700 growing degree days with the first ripening occurring around 1142 growing degree days [16].
Wild blueberries are susceptible to numerous diseases that significantly impact yield and quality [17,18]. These diseases can originate from environmental conditions or spread from nearby plants [17,18]. Identifying these diseases is crucial for preventing them in future cropping cycles. While some diseases are easily identifiable, others require expert identification [19,20]. Red Leaf disease (Exobasidium vaccinii), a common foliar disease, is characterized by red or brown leaf discoloration and significantly reduces yield [17,21]. Accurate weed identification also optimizes treatment application, reducing environmental impact and costs. Hair fescue (Festuca filiformis Pourr.), prevalent in 75% of wild blueberry fields in Nova Scotia, significantly reduces crop yield [12,22,23]. Traditional broadcast spraying of agrochemicals can be replaced with spot spraying by precisely identifying weed locations, offering environmental and economic benefits [24]. These challenges highlight the need for precise identification and management of diseases and weeds in wild blueberry fields, underscoring the importance of advanced image annotation techniques. However, existing methods for image annotation in precision agriculture are limited by their reliance on manual or semi-automated approaches, which are labor-intensive and often lack consistency in complex field environments.
Zero-shot detection methods have become pivotal in addressing the challenges in agricultural image annotation [25,26]. These methods enable the identification and segmentation of objects without pre-labeled data, which is advantageous in agriculture where manual annotation can be time consuming and costly. Mullins et al. (2024) [26] evaluated Grounding DINO and YOLO-World, on datasets relevant to wild blueberry cultivation. Grounding DINO with its strong visual representation learning and language-grounded pre-training demonstrated a superior ability to generalize to new object categories specified by textual descriptions, outperforming YOLO-World in several key metrics. YOLO-World combines the efficiency of YOLOv8′s real-time detection capabilities with advanced vision-language pre-training [27]. Across all datasets, Grounding DINO outperformed YOLO-World on metrics including Intersection over Union (IoU), precision, recall, and F1 score, making it a valuable tool for precision agriculture applications that require precise object detection [26]. By leveraging these advanced models, producers can achieve more accurate and efficient identification of diseases, weeds, and growth stages in wild blueberry fields, thereby improving crop management [26].
The task of segmentation within machine learning offers substantial potential for reducing annotation requirements through automated labeling. Unlike object detection, which uses bounding boxes, segmentation annotation requires drawing polygon segmentation masks [28,29]. The Segment Anything Model (SAM) represents a substantial advancement in computer vision, particularly in image segmentation [30,31]. SAMs were developed utilizing a combination of supervised and self-supervised learning mechanisms, achieving remarkable precision in segmenting both common and uncommon objects [30,32]. SAM’s architecture is built around a Vision Transformer (ViT) image encoder, a prompt encoder, and a lightweight mask decoder [30]. The ViT encoder generates image embeddings, while the prompt encoder processes user-provided inputs such as points, boxes, or coarse masks. These embeddings are then fused by the mask decoder to produce high-resolution segmentation outputs. This modular design allows for flexibility and adaptability across segmentation tasks, enabling the tailoring of components to specific application requirements, including agricultural contexts.
Grounded SAM is a powerful tool that integrates the capabilities of Grounding DINO and SAM to excel in both object detection and segmentation tasks [33]. The model begins with feature extraction using the network’s backbone, employing attention mechanisms to focus on relevant parts of the image, enhancing the accuracy of detection and segmentation. Grounded SAM processes images at multiple scales to capture objects of different sizes, ensuring that both small and large objects are accurately detected and segmented. Grounded SAM leverages the strengths of Grounding DINO, predicting bounding boxes and class labels for each detected object. Once objects are detected, the model applies the semantic segmentation techniques from SAM to generate detailed segmentation maps [33]. The integration of contextual information from the entire image further improves segmentation quality, allowing the model to accurately segment objects in complex scenes.
Existing methods for image annotation in precision agriculture face several limitations. Traditional manual and semi-automated approaches are labor-intensive, time-consuming, and often lack consistency, particularly in complex field environments [34,35]. While zero-shot learning models have recently demonstrated capabilities in general-purpose detection and segmentation tasks, their application to agriculture remains largely unexplored [36], with wild blueberry applications in particular being completely unexplored. There is a critical need for innovative adaptations of these advanced models to the unique challenges of agricultural environments, particularly for improving the efficiency and accuracy of instance segmentation. Zero-shot and openset detection and segmentation have seen growing applications in computer vision, including in agricultural contexts such as plant disease identification, pest detection, and crop phenotyping [37,38,39]. However, existing agricultural studies have treated zero-shot detection and segmentation as separate stages or rely on fixed prompt sets, rather than integrating dynamic prompt tuning and segmentation refinement in a unified pipeline. To date, the use of Grounding DINO and YOLO-World sequentially with SAM2 has not been evaluated in agriculture. Furthermore, no studies have been conducted using this approach to identify multiple targets in the wild blueberry cropping system.
The objective of this research was to address the challenge of efficient image annotation in precision agriculture by developing and validating an automated annotation pipeline leveraging zero-shot learning. This study integrates Grounding DINO and YOLO-World sequentially with the SAM2 segmentation framework to enhance instance segmentation annotation efficiency. The approach is tailored to identifying diverse agricultural targets within wild blueberry fields, including ripe wild blueberries (“Berry”), F2 developmental buds (“Bud”), hair fescue grass (“Fescue”), and red leaf disease infected plants (“Red leaf”). By systematically optimizing and comparing these models, this research establishes a novel, replicable pipeline for automated instance segmentation in complex field environments.
While previous work [26] evaluated zero-shot object detection frameworks on wild blueberry datasets, it focused exclusively on generating bounding box annotations. In contrast, the present study introduces a novel sequential pipeline that combines zero-shot detection (Grounding DINO and YOLO-World) with SAM2 for detailed instance segmentation. Additionally, this work includes an extensive optimization of textual prompts and confidence thresholds to maximize detection and segmentation accuracy. By integrating segmentation and refining prompt strategies, this approach advances beyond basic object detection to enable fully automated, high-resolution instance annotations suitable for complex agricultural environments.
The innovation of this work lies in several key aspects. It represents the first application of sequential zero-shot detection and segmentation models to multiple agricultural targets, addressing unique challenges such as variable lighting, foliage, hardware, and target variability. The research also introduces an optimization pipeline for prompt tuning and confidence adjustments, improving performance. Through an evaluation of interactions between Grounding DINO and YOLO World with four SAM2 models (Tiny, Small, Base, and Large), the study identifies optimal combinations that balance accuracy and efficiency. Specific objectives of this research include: (1) To apply zero-shot learning models to detect and segment multiple agricultural targets in wild blueberry fields. (2) To optimize and compare model performance, determining the most effective configurations for accurate and efficient annotation.

2. Materials and Methods

2.1. Dataset Preparation

The dataset used in this study comprised four target classes: developmental buds (F2 stage) (Figure 1A), ripe wild blueberries (Figure 1B), red leaf disease (Figure 1C), and hair fescue (Figure 1D) [26]. Image acquisition included controlled laboratory setups, ground-based field photography, and aerial drone surveys, summarized in Table 1, spanning from June of 2022 to May of 2024. The drone flight altitude of 76 m was selected to ensure sufficient ground resolution while maximizing field coverage, consistent with established imaging protocols in precision agriculture. Surveying at an altitude of 76 m resulted in a spatial resolution of 8.63 mm. Ground truth annotations, including both bounding boxes and polygon masks, were prepared using Roboflow’s polygon tool (Roboflow, Inc., Des Moines, IA, USA) and used for evaluating model performance. Representative examples for each class are shown in Figure 1.
Figure 1. Examples from each dataset with ground truth segmentation masks overlayed (magenta) where (A) shows F2 stage developmental buds, (B) shows ripe wild blueberries, (C) shows the red leaf disease that can be found in wild blueberry fields, and (D) shows hair fescue grass in a sprout stage wild blueberry field.
Table 1. Dataset Summary.

2.2. Model Selection

For the initial detection stage within the pipeline, state-of-the-art zero-shot learning models, specifically ViTs and contrastive language-image pre-training (CLIP)-based models were selected for their robust performance in zero-shot scenarios [40]. Grounding DINO, utilizing ViTs, and YOLO-World, using CLIP for encoder pretraining, were chosen for comparison. Grounding DINO’s Swin-T and Swin-B variants were studied to assess the trade-offs between computational efficiency and detection performance, with Swin-T designed for real-time processing and Swin-B offering higher capacity for better feature extraction. The YOLO-World framework included four variants (YOLO-World-s, YOLO-World-m, YOLO-World-l, YOLO-World-x) to evaluate the impact of model complexity and size on annotation performance.
Four variants of the SAM2 were used in conjunction with Grounding DINO and YOLO-World frameworks to assess their combined performance: SAM2-Tiny, SAM2-Small, SAM2-Base, and SAM2-Large. The SAM2 variants differ in model size (measured in millions of parameters), segmentation accuracy, and computational requirements. Tiny and Small are designed for faster processing and lower resource use, making them suitable for real-time or edge applications. Base provides a balance between accuracy and efficiency, while Large offers the highest segmentation precision but at increased computational cost. These characteristics are summarized in Table 2.
Table 2. SAM2 Variant Comparison adapted from Meta AI GitHub repository [41].
To provide further clarity on the underlying system structure, an architecture diagram (Figure 2) outlines the internal design of the selected models used in the annotation pipeline. The architecture diagram visualizes the individual data pathways for both image and text prompts, the feature extraction and enhancement stages, and the processes through which model outputs are generated.
Figure 2. System architecture diagram of the three core modules used in the annotation pipeline: Grounding DINO (top), SAM2 (middle), and YOLO-World (bottom). Each diagram illustrates the model-specific flow of image and prompt inputs through their respective backbones, encoders, and attention mechanisms, leading to detection or segmentation outputs.

2.3. Optimization and Tuning

To increase the performance of the zero-shot models several steps were taken to optimize the input prompts and confidences. A list of 100 potential prompts for each category was created, through a combination of human input and the GPT-4o Large Language Model (LLM) (OpenAI, San Francisco, CA, USA, 2024), followed by removal of any duplicate or inaccurate prompts of the given category. For each category, every prompt in the list was tested with a preset confidence of 0.30 for Grounding DINO Swin-T and 0.001 for YOLO-World-l. Using the ten highest mean Intersection over Union (mIoU) [30] yielding prompts, the prompts and confidence were further optimized. For each of the ten prompts, the confidence was optimized to two significant figures, if the optimal confidence was below 0.01 the process would continue to lower, to a maximum of five decimal places of resolution. The prompt and confidence pair with the highest mIoU was selected for testing in the respective category. This process was completed separately for each object detection model in each category.
For hair fescue, red leaf and buds inferenced using YOLO-World, many images exhibited classification, or bounding boxes comprising more than 90% of the image area. To alleviate this, a label cleaning algorithm was created which removes all bounding boxes over a specified area threshold, exceptions were made with images only containing one bounding box. The box threshold was set to 0.4 for berries and red leaf, 0.5 for fescue, and 0.25 for buds. For fescue, the high-resolution drone images were tiled into 8 × 8 chunks, resulting in image dimensions is 1024 × 682, using the preprocessing option in Roboflow’s dataset generation, as the full-size images (8192 × 5460 pixels) are too difficult for the zero-shot models to discern target boundaries effectively due to the relative size and target density [26]. The text threshold, the minimum confidence for a token to be considered a detection in Grounding DINO, was set to 0.1 and remained unchanged.
A detailed flowchart of the proposed annotation and optimization pipeline, summarizing data preprocessing, prompt generation, iterative confidence tuning, zero-shot object detection, segmentation integration, and statistical evaluation, is provided in Figure 3 to illustrate the full methodological sequence.
Figure 3. Overview of the automated annotation and optimization pipeline integrating zero-shot object detection, LLM-guided prompt generation, confidence tuning, and SAM segmentation, followed by statistical performance evaluation.

2.4. SAM Automatic Annotation Integration

Bounding boxes were passed into SAM2, so that only the area within each box was segmented. Grounding DINO outputs boxes in a format where the normalized x-position, y-position, width and height of each box is listed in the specified order. It was necessary to convert these boxes to the format accepted by SAM2 which consists of the x and y coordinates (in pixels) of the top-left corner of the box, followed by the x and y-coordinates of the bottom-right corner.

2.5. Evaluation Procedure

Both the zero-shot detection and segmentation models were deployed without additional training. Each model was evaluated on an Alienware Aurora R11 desktop computer (Dell Inc., Round Rock, TX, USA) with 3.7 GHz 10-core Intel Core i9-10900K CPU (Intel Corporation, Santa Clara, CA, USA), 128 GB of DDR4-3200 MT/s RAM and an Nvidia GeForce RTX 3090 GPU (Nvidia Corp., Santa Clara, CA, USA). Grounding DINO, YOLO-World, and SAM2 models were assessed on the same datasets to ensure consistent evaluation. IoU and processing time were calculated and recorded for each image within each dataset. The IoU is crucial in object detection and instance segmentation tasks, as it quantifies the overlap between the predicted and actual bounding boxes/polygons, providing a spatial accuracy assessment [42].

2.6. Statistical Analysis

After conducting the evaluations, the performance metrics were subjected to statistical analysis to determine any significant differences between the detection model and SAM2 combinations. A 6x4 factorial design was employed, involving two primary factors: detection model, with six levels (Grounding DINO Swin-T, Grounding DINO Swin-B, YOLO-World-s, YOLO-World-m, YOLO-World-l, YOLO-World-x), and SAM2, with four levels (Tiny, Small, Base, Large). Each combination of these factors was tested with six replicates, resulting in a total of 144 observations. The replicates were selected using simple random sampling, creating subsets of the entire dataset. The images used for each replicate were the same across model combinations. The observations were analyzed through an ANOVA general linear model to assess the main and interaction effects on segmentation performance metrics of mIoU and processing time [43]. Following the ANOVA, multiple mean comparisons were performed using Tukey’s Honest Significant Difference (HSD) test using Minitab 21 (Minitab, LLC., Minitab 21, State College, PA, USA) statistical software with an acceptance level α of 0.05 [43]. This approach allowed for the identification of significant differences between the means of the performance metrics for different model combinations. In addition to the statistical analyses, eta squared (η2) was recorded as a measure of effect size for the ANOVA analyses. The use of CIELAB color space’ perceptual lightness (L*) image brightness metric was calculated at the pixel level for each image in each category to find the average L* and deviation of L* across images in each category.

3. Results and Discussion

The preliminary optimization of the zero-shot detection models found Grounding DINO and YOLO-World models preferred different text inputs for label optimization (Table 3). This was similar in confidence threshold optimization where Grounding DINO Swin-T typically achieved a higher confidence threshold compared to Swin-B. Similarly, YOLO-World models demonstrated varying confidence thresholds (Table 3).
Table 3. Summary of zero-shot detection mechanism optimization of confidence and prompt inputs across datasets and models.
The variation in textual prompts and confidence thresholds was determined through empirical optimization to maximize mIoU for each category. Larger models such as Swin-T and Swin-B could leverage more complex relationships with prompts due to their advanced feature extraction and higher capacity, enabling precise differentiation of similar targets. In contrast, YOLO-World models required extremely low confidence thresholds (0.00–4.00%) to ensure sufficient object detection coverage, reflecting their design for real-time efficiency rather than nuanced language-vision alignment. These differences highlight the impact of model architecture on prompt sensitivity and confidence calibration during open-vocabulary annotation tasks.
The optimized prompts and confidences were subsequently utilized for detection boxes supplied to the SAM2 model. The results of the bud detection indicated that the Swin-B model, when paired with any SAM2 model, achieved the highest performance in terms of mIoU (Table 4). Although differences among SAM2 models were minimal, a significant difference was noted in the zero-shot detection models. In these cases, each SAM2 model, when used sequentially with the Grounding DINO detection models, exhibited superior performance (mIoU > 0.641) compared to the YOLO-World models (mIoU (excluding model YOLO-World-x) < 0.479). The analysis of variance (ANOVA) using a general linear model revealed that the SAM2 model had no significant effect (p = 0.986), whereas the detection model had a significant effect (p < 0.001). Furthermore, there was no significant interaction effect between the SAM2 models and the detection models (p = 1.000).
Table 4. Mean Intersection over Union (mIoU) performance between SAM2 models and zero-shot detection model variants on developmental wild blueberry bud images (α = 0.05) (L* = 60.04 ± 19.65) (maximum η2 = 0.31).
From Table 4, the recommended models for developmental bud stage segmentation were the Grounding DINO with Swin-B and any SAM2 model. The Swin-B model consistently achieved the highest mIoU across various SAM2 model configurations, making it suitable for accurate segmentation. Given the statistically insignificant performance differences between SAM2 models (p = 0.986), practical considerations like computational resources and processing efficiency become crucial. The SAM2-Tiny model, despite being lightweight, does not compromise performance and is recommended for reducing compute requirements and processing time.
When evaluated on ripe wild blueberry images, the SAM2-Tiny model achieved mIoU values ranging from 0.631 to 0.731 (Table 5), with the highest performance observed in the Swin-B variant at 0.731 (±0.200). The SAM2-Small model also showed strong performance, achieving the highest mIoU with the Swin-B variant at 0.738 (±0.189). SAM2-Small exhibited lower variability across detection models compared to SAM2-Large, suggesting greater consistency and reliability. This makes SAM2-Small a strong candidate for applications requiring stability. The SAM2-Base model demonstrated mIoU values from 0.631 to 0.723, with its peak performance also with the Swin-B variant at 0.723 (±0.202). Both SAM2-Tiny and SAM2-Small consistently outperformed other models, indicating superior segmentation performance for ripe blueberries. The SAM2-Large model reached its highest mIoU with the Swin-B variant at 0.738 (±0.194). Despite the differences in mIoU values (~10%), the large standard deviations resulted in the ANOVA general linear model indicating that the main effects and interaction effects were not significant (p > 0.594).
Table 5. Mean Intersection over Union (mIoU) performance between SAM2 models and zero-shot detection model variants on ripe wild blueberries images (α = 0.05) (L* = 47.53 ± 19.19) (maximum η2 = 0.03).
All SAM2 model variants (Tiny, Small, Base, and Large) demonstrated effective segmentation capabilities, as indicated by their mIoU values. The Swin-B zero-shot detection model variant significantly consistently outperformed other model sizes across all SAM2 variants, emphasizing its robust feature extraction and detection capabilities, which are crucial for optimal segmentation of ripe wild blueberries. Although the YOLO-World model variants demonstrated incremental improvements in mIoU values with increasing model sizes, these differences were not statistically significant. This indicates that scaling up YOLO-World models results in marginal performance gains. Therefore, the selection of YOLO-World model sizes should be guided by factors such as computational efficiency and resource availability rather than expectations of significant improvements in segmentation performance.
When evaluated on images depicting red leaf disease, the SAM2-Tiny model exhibited mIoU values ranging from 0.245 to 0.884, with its highest performance observed with the Swin-T model variant, achieving an mIoU of 0.884 (±0.110) (Table 6). There was considerable variation in performance across different zero-shot detection model sizes, suggesting that the choice of detection model significantly impacts the segmentation accuracy. The SAM2-Small model performed similarly, with mIoU values ranging from 0.280 to 0.885, reaching its peak with the YOLO-World-x variant at 0.885 (±0.097), indicating a strong ability to handle red leaf disease segmentation consistently across different detection models. The SAM2-Base model showed mIoU values from 0.223 to 0.843, with the best performance recorded with the Swin-B variant at an mIoU of 0.843 (±0.212). This model, like SAM2-Tiny, demonstrated significant performance variation depending on the zero-shot detection model used. The SAM2-Large model achieved mIoU values ranging from 0.351 to 0.905, with its highest mIoU observed with the Swin-T variant at 0.905 (±0.114), underscoring its enhanced capacity to manage complex segmentation tasks. Across all SAM2 variants, the Swin-T and Swin-B detection models consistently resulted in the highest mIoU scores, emphasizing their robust feature extraction capabilities.
Table 6. Mean Intersection over Union (mIoU) performance between SAM2 models and zero-shot detection model variants on red leaf disease images (α = 0.05) (L* = 60.26 ± 19.55) (maximum η2 = 0.46).
The broader range of mIoU values for models applied to red leaf disease images compared to ripe wild blueberries suggests that segmenting diseased leaf areas may present additional challenges. Despite notable disparities in mIoU values, large standard deviations across different model combinations resulted in no significant differences identified. ANOVA results indicated that the detection model had a significant effect on mIoU performance (p < 0.001), implying that the choice of zero-shot detection model significantly impacts segmentation accuracy for red leaf disease images. In contrast, neither the SAM2 model variant nor the interaction effects significantly affected mIoU performance (p > 0.421).
This analysis was crucial for evaluating the effectiveness of various model configurations in accurately segmenting diseased leaf areas, a key aspect of precision agriculture and plant disease management. Notably, the Swin-T and Swin-B zero-shot detection models consistently achieved the highest mIoU scores across all SAM2 variants with the exception of YOLO-World-x in combination with SAM2-Small which achieved the highest. The consistently high mIoU scores demonstrated robust feature extraction and detection capabilities ideal for segmenting red leaf disease. The Swin-T model variant achieved the highest mIoU values for the Grounding DINO framework for SAM2-Tiny (0.884 ± 0.110), SAM2-Small (0.878 ± 0.112), and SAM2-Large (0.905 ± 0.114), showcasing its effectiveness across different configurations. Similarly, the Swin-B model performed exceptionally well, with high mIoU scores for SAM2-Tiny (0.866 ± 0.135), SAM2-Small (0.867 ± 0.130), and SAM2-Large (0.887 ± 0.131), and the highest SAM2-Base score (0.843 ± 0.212).
In contrast, the YOLO-World models exhibited more variable performance, with mIoU values differing significantly across SAM2 variants. YOLO-World-m, in particular, showed relatively low mIoU values across all SAM2 variants, indicating less reliability for segmenting red leaf disease compared to Swin models. Given the non-significant performance differences among SAM2 models, the choice of SAM2 variant can be optimized based on computational efficiency and resource availability. The high variability in mIoU performance among YOLO-World model variants underscores the importance of careful selection and validation of detection models in practical applications. While some configurations may achieve high accuracy, others may perform poorly, leading to inconsistent results in real-world settings.
The results of the hair fescue segmentation indicated that the Swin-T model achieved the highest mIoU (0.694), significantly outperforming the other models across all SAM2 variants, with the SAM2-Large model achieving the peak performance of 0.694 (±0.175) (Table 7). This indicates that the Swin-T model is particularly adept at differentiating hair fescue from the surrounding blueberry plants in aerial imagery. The Swin-B model also showed strong performance, achieving its highest mIoU score of 0.647 (±0.223) with the SAM2-Large model. These results indicate that both Swin-T and Swin-B models are robust and effective for high-accuracy segmentation tasks in complex agricultural environments. In contrast, the YOLO-World models performed poorly across all SAM2 variants, with significantly lower mIoU scores. The highest mIoU within the YOLO-World framework was 0.127 (±0.083) for the YOLO-World-m model paired with the SAM2-Tiny variant. This consistent underperformance suggests that the YOLO-World models are not suitable for the precise segmentation tasks required in precision agriculture, likely due to their inability to effectively capture and differentiate the intricate features present in aerial images of hair fescue in blueberry fields.
Table 7. Mean Intersection over Union (mIoU) performance between SAM2 models and zero-shot detection model variants on aerial fescue grass images (α = 0.05) (L* = 44.25 ± 7.36) (maximum η2 = 0.77).
The nature of the drone images used in this study substantially impacts model performance. Aerial imagery presents several challenges for segmentation models: the images contain complex backgrounds with varying textures and colors, making it difficult for models to distinguish between fescue and surrounding vegetation or soil; variations in lighting and the presence of shadows can create additional noise, complicating the segmentation task; drone images cover large areas with high resolution, capturing both macro and micro-level details. The Swin-T model’s superior performance suggests it effectively manages these challenges, including resolution, varying light, and complex backgrounds. However, the substantial decrease in mIoU performance from bounding box-level assessment to polygon assessment indicates that the SAM2 models are a more erroneous component in the pipeline, showing more than a 10% mIoU loss. In contrast, the YOLO-World models produced inadequate segmentation results, rendering them unsuitable for practical agricultural applications.
Across all datasets, the Swin-T and Swin-B models consistently outperformed the YOLO-World models. The Swin-T model exhibited the highest mIoU scores across different datasets, demonstrating superior feature extraction and detection capabilities. This trend was most evident in the diseased plant dataset, where the Swin-T model achieved an mIoU of 0.905 when paired with the SAM2-Large variant. This indicated the model’s exceptional ability to handle complex image features associated with disease symptoms, outperforming other datasets where the Swin-T model was still robust, but had lower mIoU scores. In comparing the developmental buds and ripe blueberries datasets, the performance of Swin-T and Swin-B models remained consistently high. However, Swin-B achieved the highest mIoU scores, slightly higher in the developmental buds (0.751) compared to the ripe blueberries dataset (0.738). This marginal difference suggests that the Swin-B model is highly effective across different stages of crop maturity. The Swin-T model also performed well in these datasets, though it was consistently behind the Swin-B model. The fescue dataset provided another perspective on model performance, where the Swin-T model led in mIoU scores. However, the performance gap between Swin-T and Swin-B was larger, indicating that the fescue dataset might present more complexity in features compared to ripe blueberries or developmental buds, allowing Swin-B to perform relatively lower to Swin-T. The YOLO-World models consistently underperformed across all datasets, showing significantly lower and more variable mIoU scores. Despite slight improvements with increased model size, the YOLO-World models failed to achieve competitive performance, emphasizing the need for more robust detection models like Swin-T and Swin-B. Additional segmentation performance metrics—including F1-score, precision, and recall for every detection–segmentation model pairing—are provided in Appendix A (Table A1, Table A2, Table A3 and Table A4) to complement the mIoU results summarized in Table 4, Table 5, Table 6 and Table 7.
An overarching finding across dataset comparisons was the choice of SAM2 model having an insignificant impact on segmentation performance when paired with the high-performing Swin-T and Swin-B models, suggesting that practical considerations such as computational efficiency and resource availability should guide their selection. This consistency across datasets implies that SAM2 models are reliably interchangeable when used with effective detection models, further simplifying the decision-making process for model deployment.
Given the insignificant differences in segmentation performance of the SAM2 model component, examining processing time helps identify the optimal combination of SAM2 and zero-shot detection models (Table 8). SAM2-Tiny and Small had the shortest processing times, with Swin-T at 0.30 ± 0.10 s and Swin-B at 0.35 ± 0.11 s and 0.36 ± 0.10 s, indicating high efficiency. In contrast, SAM2-Tiny paired with YOLO-World-x required more time, averaging 0.64 ± 0.11 s. SAM2-Base exhibited similar processing times with Grounding DINO model, where Swin-T took 0.33 ± 0.10 s and Swin-B took 0.38 ± 0.10 s. The longest time within the SAM2-Base group was for YOLO-World-l, at 1.03 ± 0.57 s. SAM2-Large had the most extended processing times, ranging from 0.42 ± 0.11 s for Swin-T to 1.23 ± 0.54 s for YOLO-World-l. Notably, all YOLO-World variants within the SAM2-Large group exceeded 0.8 s per image, indicating higher computational demands. These results suggest that SAM2-Tiny or Small, especially when paired with Swin-T or Swin-B, offers the best balance of efficiency and performance. For applications where processing time is critical, Swin-T and Swin-B models with SAM2-Tiny or Small are optimal choices, providing effective segmentation with minimal computational overhead.
Table 8. Processing time performance between SAM2 models and zero-shot detection model variants across all datasets (α = 0.05) (maximum η2 = 0.46).
The results from Table 8 show a clear distinction in processing times among the SAM2 models and zero-shot detection variants. SAM2-Tiny and Small, particularly when paired with Swin-T and Swin-B, demonstrated the shortest processing times, highlighting their computational efficiency, which is beneficial for real-time applications where speed is crucial. SAM2-Large showed the longest processing times across all model variants, with particularly high values for all YOLO-World models, suggesting that SAM2-Large, despite potentially offering higher segmentation accuracy, imposes a significant computational burden that may limit its practicality in time-sensitive applications. The differences in processing times emphasize the importance of selecting model combinations based on specific operational needs. For applications requiring rapid processing, SAM2-Tiny or SAM2-Small with Swin-T or Swin-B is recommended. Conversely, for scenarios where processing time is less critical, but accuracy is paramount, SAM2-Large with higher-performing detection models may be appropriate.

3.1. Qualitative Visualization and Prompt-Specific Observations

Bounding box detections and corresponding SAM-based segmentations for developmental buds, ripe wild blueberries, and leaf regions reveal distinct differences in detection granularity and segmentation precision across model configurations. For developmental buds (Figure 4), the Swin-B configuration consistently produced a higher count of precise detections, effectively identifying individual buds along the stem while maintaining tight bounding box alignment. In contrast, Swin-T tended to under-detect smaller buds and occasionally generalized by grouping nearby buds within broader regions. The YOLO-World models demonstrated size-dependent behavior, where the smaller configuration (YOLO-World-s) occasionally fragmented clusters, while YOLO-World-m and YOLO-World-l variants included surrounding branch structures. These results indicate that transformer-based backbones (Swin) prioritize finer detail extraction, whereas YOLO-World configurations leverage broader spatial detection, influencing detection comprehensiveness. All models occasionally identified leaf buds in addition to blueberry developmental buds, suggesting shared visual cues among morphological stages.
Figure 4. Bounding box detections for developmental buds using Swin-B and Swin-T in yellow, and YOLO-World-s, YOLO-World-m, YOLO-World-l, and YOLO-World-x configurations in blue compared to ground truth in green, illustrating differences in detection precision and segmentation outcomes.
For ripe wild blueberries (Figure 5), the Swin-B model again demonstrated strong localization accuracy, capturing most individual berries with minimal overlap or misclassification. Swin-T achieved acceptable coverage but with slightly fewer total detections, particularly in dense clusters where multiple berries overlapped. The YOLO-World configurations showed a similar trend, with the smaller variants (YOLO-World-s and YOLO-World-m) detecting fewer instances and the YOLO-World-l increasing recall but occasionally merging nearby fruits under single bounding regions. The comparison highlights that while transformer-based detectors emphasize precise regional boundaries, YOLO-World variants optimized for broader coverage at varying confidence levels.
Figure 5. Bounding box detections for ripe wild blueberries using Swin-B, Swin-T, YOLO-World-s, YOLO-World-m, YOLO-World-l, and YOLO-World-x configurations compared to ground truth, illustrating differences in detection precision and segmentation outcomes.
For leaf segmentation overlays (Figure 6), differences between model–SAM pairings became visually apparent. Both Swin-B and T integrated with SAM2 Large demonstrated superior spatial adherence to leaf boundaries, achieving high fidelity along irregular margins. The YOLO-World models displayed a wider performance range: YOLO-World-s and YOLO-World-m produced sparse masks lacking full contour closure, while YOLO-World-l increased boundary completeness but at the cost of including non-red leaf structures, whereas YOLO-World-x was comparable to the Swin-B and T combinations. These qualitative results align with mIoU findings, confirming that larger detection–segmentation pairs yield more coherent segmentation masks.
Figure 6. Overlay of segmentation masks (teal) from Swin-B, Swin-T, YOLO-World-s, YOLO-World-m, YOLO-World-l, YOLO-World-x, and ground truth, showing spatial alignment and segmentation accuracy differences.
To further assess prompt-specific sensitivity, all models were evaluated using identical textual prompts and previously optimized confidence thresholds (Figure 7). Under uniform conditions, Swin-B captured nearly all visible buds but also introduced false positives by extending detections to nearby leaf buds. Swin-T produced fewer but more selective detections, avoiding false inclusions yet missing several small targets. YOLO-World-s and YOLO-World-m demonstrated the highest selectivity but lowest completeness, while YOLO-World-l and YOLO-World-x configurations expanded recall at variable precision. This suggests that transformer-based models leverage prompts for broader contextual comprehension, whereas YOLO-World relies more heavily on localized salience, emphasizing the trade-off between generalization and focus.
Figure 7. Qualitative comparison of Swin-B (yellow), Swin-T (purple), and YOLO-World models (red) using the same prompt at previously optimized confidence level, demonstrating model-specific differences in detection granularity and region selection.

3.2. Failure Analysis

Across the four representative examples, several consistent failure behaviors were observed among the detection–segmentation pipelines relative to the ground truth (Figure 8). The qualitative results highlight clear and recurrent failure tendencies across the different model families. In the developmental bud and ripe blueberry images, YOLO-World variants frequently produced coarse detections that merged objects with surrounding vegetation and overlooked smaller or partially occluded instances. Swin-T demonstrated the opposite behavior, generating clean but conservative segmentations that missed valid targets, while Swin-B improved overall coverage yet still failed to recover peripheral objects and occasionally extended beyond true object boundaries. These patterns reflect consistent model-specific biases, with YOLO-World relying heavily on broad contextual cues and the Swin-based detectors emphasizing localized structure with varying degrees of completeness.
Figure 8. Qualitative comparison of model outputs across four representative tasks: developmental buds, ripe blueberries, red leaf disease, and aerial hair fescue. Columns show predictions from the top performing YOLO-World model (red/blue) (left), Swin-T (magenta) (center-left), Swin-B (yellow) (center-right), and the ground truth (green) (right).
The red leaf disease and aerial hair fescue scenes in Figure 8 further illustrate these differences. YOLO-World again produced oversized detections driven by global chromatic or textural contrast rather than the detailed morphology of the target regions, resulting in substantial over-segmentation. Swin-T and Swin-B localized symptomatic and weedy areas more accurately but continued to struggle with irregular shapes, sometimes introducing small false positives (Swin-T) or fragmenting larger structures into multiple partial detections (Swin-B). These qualitative failures align with the quantitative mIoU trends and the difficulty of achieving reliable fine-scale delineation in agricultural imagery, where dense vegetation, occlusion, and heterogeneous backgrounds pose persistent challenges for both detection and segmentation.
These qualitative analyses provide clear visual evidence of model strengths and limitations, support the quantitative findings, and inform practical selection of model–prompt configurations for various agricultural applications.

4. Conclusions

This study integrated zero-shot detection models and SAM2 for automated image annotation, focusing on identifying ripe wild blueberries, F2 developmental buds, hair fescue, and red leaves in wild blueberry fields. It evaluated the performance of Grounding DINO and YOLO-World models with SAM2 to develop an efficient instance segmentation framework. The objectives included applying zero-shot learning for target detection, assessing annotation accuracy, and optimizing model combinations for precision and efficiency.
This study found that Swin-T and Swin-B zero-shot detection models consistently achieved the highest mIoU scores across all SAM2 variants for wild blueberry buds, ripe blueberries, red leaf disease, and hair fescue. SAM2-Large paired with Swin-T achieved the highest mIoU for hair fescue (0.694 ± 0.175) and red leaf disease (0.905 ± 0.114), while Swin-B with SAM2-Large performed best for wild blueberry buds (0.751 ± 0.154). For ripe blueberries, Swin-B paired with SAM2-Small achieved the highest mIoU (0.738 ± 0.189), excelling in complex segmentation. Regarding processing time, SAM2-Tiny, Small, and Base demonstrated the fastest performance with Swin-T (0.30–0.33 s) and Swin-B (0.35–0.38 s). Although SAM2-Large remains within acceptable limits for some real-time applications (0.42–0.48 s with Swin-T and B models), its higher computational demand makes it less practical when in comparison to Small and Tiny. These results highlight the effectiveness of Swin-T and Swin-B models with SAM2 variants for segmenting agricultural targets, particularly when computational efficiency is critical. SAM2-Tiny and Small offer a balance between accuracy and speed, making them ideal for real-time applications, while the choice of SAM2 model can be tailored to operational needs. However, the study’s findings are limited by the specificity of the datasets to particular geographic locations and conditions, which may affect generalizability. Although zero-shot models are designed for broad applicability, performance may vary with different textual descriptions or unseen object categories.
Future research should aim to integrate zero-shot mechanisms into a fully automated annotation pipeline for the wild blueberry cropping system, minimizing manual annotation efforts. Priorities include optimizing Grounding DINO for complex agricultural scenarios, conducting extensive real-world testing across diverse environments to enhance robustness and generalization, and expanding the approach to other agricultural datasets to assess its broader applicability. Exploring additional zero-shot models to improve annotation accuracy and efficiency is also essential. These advancements could significantly enhance agricultural image annotation, improving crop management practices in wild blueberry cultivation and beyond.

Author Contributions

C.C.M.: Conceptualization, data curation, formal analysis, investigation, methodology, software, visualization, writing—original draft. T.J.E.: Funding acquisition, project administration, resources, supervision, writing—review and editing. R.J.: Methodology, software, visualization, writing—original draft. C.L.T.: Data curation, writing—review and editing, writing—original draft. P.J.H.: Data curation, writing–review and editing. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by Natural Sciences and Engineering Research Council (NSERC) of Canada Discovery Grants Program (RGPIN-06295-2019) and Alliance Grant (ALLRP 588581-23).

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author.

Acknowledgments

The authors thank Bragg Lumber Company and landowners in Nova Scotia and New Brunswick for using commercially managed wild blueberry fields. They also acknowledge the efforts of Dalhousie’s Agricultural Mechanized Systems Research Lab and Atlantic Institute for Digital Agriculture. The authors declare that generative AI was used solely in the scope of data collection for the research itself and was not used for the writing of the article.

Conflicts of Interest

The authors declare that they have no competing financial or personal interests that could influence the work reported in this study.

Abbreviations

The following abbreviations are used in this manuscript:
AIArtificial Intelligence
ANOVAAnalysis of Variance
CLIPContrastive Language–Image Pretraining
CPUCentral Processing Unit
GPUGraphics Processing Unit
HSDHonest Significant Difference
IoUIntersection over Union
LLMLarge Language Model
mIoUMean Intersection over Union
SAMSegment Anything Model
SAM2Segment Anything Model Version 2
ViTVision Transformer
YOLOYou Only Look Once

Appendix A

Table A1. F1-score, precision, and recall performance between SAM2 models and zero-shot detection model variants on aerial fescue grass images.
Table A1. F1-score, precision, and recall performance between SAM2 models and zero-shot detection model variants on aerial fescue grass images.
Detection ModelSegmentation ModelF1-ScorePrecisionRecall
YOLO-World-sTiny0.198 ± 0.1110.1144 ± 0.07380.9993 ± 0.001
YOLO-World-mTiny0.2165 ± 0.11910.1276 ± 0.07990.9838 ± 0.049
YOLO-World-lTiny0.198 ± 0.1110.1144 ± 0.07380.9993 ± 0.001
YOLO-World-xTiny0.198 ± 0.1110.1144 ± 0.07380.9993 ± 0.001
YOLO-World-sSmall0.198 ± 0.1110.1144 ± 0.07380.9993 ± 0.001
YOLO-World-mSmall0.2151 ± 0.11750.1257 ± 0.07770.9971 ± 0.007
YOLO-World-lSmall0.198 ± 0.1110.1144 ± 0.07380.9993 ± 0.001
YOLO-World-xSmall0.198 ± 0.1110.1144 ± 0.07380.9993 ± 0.001
YOLO-World-sBase0.194 ± 0.11560.1123 ± 0.0760.9527 ± 0.1465
YOLO-World-mBase0.194 ± 0.11560.1123 ± 0.0760.9527 ± 0.1465
YOLO-World-lBase0.1981 ± 0.1110.1145 ± 0.07380.9993 ± 0.001
YOLO-World-xBase0.1937 ± 0.11580.1122 ± 0.0760.9507 ± 0.1494
YOLO-World-sLarge0.1927 ± 0.11640.1117 ± 0.07640.9424 ± 0.1513
YOLO-World-mLarge0.1928 ± 0.11630.1117 ± 0.07640.9431 ± 0.1512
YOLO-World-lLarge0.1982 ± 0.11030.1103 ± 0.07350.9961 ± 0.01
YOLO-World-xLarge0.192 ± 0.11680.1113 ± 0.07660.9369 ± 0.1538
SwinBTiny0.7295 ± 0.17640.6543 ± 0.23110.9177 ± 0.1281
SwinBSmall0.7414 ± 0.18360.6637 ± 0.22830.9312 ± 0.1317
SwinBBase0.7576 ± 0.18020.6954 ± 0.22820.917 ± 0.1302
SwinBLarge0.7617 ± 0.17440.711 ± 0.22630.9037 ± 0.1368
SwinTTiny0.767 ± 0.15790.7423 ± 0.19020.8345 ± 0.1611
SwinTSmall0.7716 ± 0.16760.7401 ± 0.20220.8599 ± 0.1536
SwinTBase0.79 ± 0.16660.7881 ± 0.19380.8449 ± 0.1739
SwinTLarge0.8173 ± 0.1020.8206 ± 0.10090.8333 ± 0.1528
Table A2. F1-score, precision, and recall performance between SAM2 models and zero-shot detection model variants on red leaf disease images.
Table A2. F1-score, precision, and recall performance between SAM2 models and zero-shot detection model variants on red leaf disease images.
Detection ModelSegmentation ModelF1-ScorePrecisionRecall
YOLO-World-sTiny0.5331 ± 0.38040.482 ± 0.41320.9759 ± 0.0312
YOLO-World-mTiny0.331 ± 0.27970.3277 ± 0.3340.7925 ± 0.315
YOLO-World-lTiny0.3392 ± 0.38480.3218 ± 0.41720.6648 ± 0.4397
YOLO-World-xTiny0.8994 ± 0.11270.9322 ± 0.15090.8944 ± 0.1162
YOLO-World-sSmall0.486 ± 0.38750.4465 ± 0.43260.9772 ± 0.0506
YOLO-World-mSmall0.3481 ± 0.32790.4145 ± 0.40820.8303 ± 0.3282
YOLO-World-lSmall0.3346 ± 0.37950.3174 ± 0.41140.6606 ± 0.4382
YOLO-World-xSmall0.9364 ± 0.05510.9636 ± 0.04880.9418 ± 0.082
YOLO-World-sBase0.5463 ± 0.38940.5227 ± 0.43590.9568 ± 0.0909
YOLO-World-mBase0.4139 ±0.33630.3496 ± 0.35850.9744 ± 0.0504
YOLO-World-lBase0.2738 ± 0.34460.3099 ± 0.42060.507 ± 0.459
YOLO-World-xBase0.8735 ± 0.17730.9175 ± 0.16860.8899 ± 0.1989
YOLO-World-sLarge0.4807 ± 0.39520.5252 ± 0.44020.8948 ± 0.2709
YOLO-World-mLarge0.4302 ± 0.3510.3557 ± 0.35330.9927 ± 0.0135
YOLO-World-lLarge0.325 ± 0.40090.3087 ± 0.41830.5739 ± 0.4709
YOLO-World-xLarge0.9011 ± 0.17590.9016 ± 0.2240.9489 ± 0.04
SwinBTiny0.9281 ± 0.08050.9824 ± 0.0140.8885 ± 0.129
SwinBSmall0.9252 ± 0.07790.9724 ± 0.04610.8927 ± 0.1248
SwinBBase0.9368 ± 0.07690.9818 ± 0.00920.9045 ± 0.1264
SwinBLarge0.9391 ± 0.07710.9779 ± 0.12530.909 ± 0.1177
SwinTTiny0.9401 ± 0.06420.9795 ± 0.10370.9097 ± 0.1029
SwinTSmall0.9331 ± 0.06480.9682 ± 0.04830.9074 ± 0.101
SwinTBase0.9377 ± 0.07590.9828 ± 0.0080.9055 ± 0.1256
SwinTLarge0.953 ± 0.05880.9775 ± 0.02650.9342 ± 0.0929
Table A3. F1-score, precision, and recall performance between SAM2 models and zero-shot detection model variants on developmental wild blueberry bud images.
Table A3. F1-score, precision, and recall performance between SAM2 models and zero-shot detection model variants on developmental wild blueberry bud images.
Detection ModelSegmentation ModelF1-ScorePrecisionRecall
YOLO-World-sTiny0.6064 ± 0.21470.6532 ± 0.19960.6289 ± 0.2835
YOLO-World-mTiny0.5239 ± 0.16290.4935 ± 0.25610.7784 ± 0.234
YOLO-World-lTiny0.6774 ± 0.13010.6021 ± 0.18030.8486 ± 0.1659
YOLO-World-xTiny0.7441 ± 0.12710.7454 ± 0.19370.8187 ± 0.1808
YOLO-World-sSmall0.6109 ± 0.21840.7181 ± 0.16990.6103 ± 0.2861
YOLO-World-mSmall0.5206 ± 0.10430.5172 ± 0.25670.7896 ± 0.2671
YOLO-World-lSmall0.6697 ± 0.13950.5919 ± 0.19440.8559 ± 0.1593
YOLO-World-xSmall0.7316 ± 0.14010.738 ± 0.21190.8064 ± 0.1775
YOLO-World-sBase0.5711 ± 0.25650.6826 ± 0.24640.5207 ± 0.2795
YOLO-World-mBase0.5025 ± 0.12220.4849 ± 0.24320.7622 ± 0.2629
YOLO-World-lBase0.7255 ± 0.12060.7101 ± 0.18630.7933 ± 0.1541
YOLO-World-xBase0.7181 ± 0.15380.8042 ± 0.18750.714 ± 0.2103
YOLO-World-sLarge0.6078 ± 0.2140.7682 ± 0.21970.5255 ± 0.2368
YOLO-World-mLarge0.5224 ± 0.16420.5008 ± 0.26380.7717 ± 0.2584
YOLO-World-lLarge0.7597 ± 0.09540.7762 ± 0.19620.8208 ± 0.1615
YOLO-World-xLarge0.7206 ± 0.15460.8182 ± 0.19540.7122 ± 0.2099
SwinBTiny0.8529 ± 0.11260.8948 ± 0.03550.8426 ± 0.19
SwinBSmall0.8542 ± 0.11280.8989 ± 0.0390.8421 ± 0.1903
SwinBBase0.8564 ± 0.11560.8994 ± 0.04430.8443 ± 0.1898
SwinBLarge0.8612 ± 0.11670.9165 ± 0.04410.8386 ± 0.1884
SwinTTiny0.7622 ± 0.16540.917 ± 0.05830.6764 ± 0.2079
SwinTSmall0.7645 ± 0.1650.9235 ± 0.06190.677 ± 0.2078
SwinTBase0.7673 ± 0.16690.9253 ± 0.05960.6781 ± 0.207
SwinTLarge0.7698 ± 0.16810.9405 ± 0.06240.6743 ± 0.206
Table A4. F1-score, precision, and recall performance between SAM2 models and zero-shot detection model variants on ripe wild blueberry images.
Table A4. F1-score, precision, and recall performance between SAM2 models and zero-shot detection model variants on ripe wild blueberry images.
Detection ModelSegmentation ModelF1-ScorePrecisionRecall
YOLO-World-sTiny0.7639 ± 0.11970.8295 ± 0.06120.7251 ± 0.1815
YOLO-World-mTiny0.8276 ± 0.11130.839 ± 0.08050.8284 ± 0.155
YOLO-World-lTiny0.8108 ± 0.14290.7779 ± 0.15850.8497 ± 0.1264
YOLO-World-xTiny0.8029 ± 0.12740.7667 ± 0.12760.8591 ± 0.1585
YOLO-World-sSmall0.7599 ± 0.11920.82 ± 0.09880.7421 ± 0.1957
YOLO-World-mSmall0.818 ± 0.110.8428 ± 0.08950.8057 ± 0.1458
YOLO-World-lSmall0.8106 ± 0.14390.7844 ± 0.16060.8417 ± 0.1287
YOLO-World-xSmall0.7984 ± 0.12210.768 ± 0.12740.8491 ± 0.1543
YOLO-World-sBase0.7649 ± 0.11050.8444 ± 0.05690.7238 ± 0.1823
YOLO-World-mBase0.8174 ± 0.1080.8337 ± 0.07940.8118 ± 0.1471
YOLO-World-lBase0.8096 ± 0.14240.7762 ± 0.15870.8494 ± 0.1262
YOLO-World-xBase0.7944 ± 0.12580.7577 ± 0.13040.852 ± 0.1558
YOLO-World-sLarge0.783 ± 0.12550.8631 ± 0.06370.7401 ± 0.192
YOLO-World-mLarge0.8169 ± 0.11280.8384 ± 0.09660.808 ± 0.1458
YOLO-World-lLarge0.8121 ± 0.14020.7833 ± 0.15710.8462 ± 0.1245
YOLO-World-xLarge0.8027 ± 0.12460.7728 ± 0.12590.8507 ± 0.1555
SwinBTiny0.8274 ± 0.1480.7973 ± 0.19150.885 ± 0.0996
SwinBSmall0.8364 ± 0.13870.8147 ± 0.18340.8811 ± 0.0951
SwinBBase0.833 ± 0.13760.8041 ± 0.1770.8836 ± 0.0984
SwinBLarge0.8405 ± 0.13460.8183 ± 0.17670.8832 ± 0.0959
SwinTTiny0.8305 ± 0.12420.8941 ± 0.08260.7865 ± 0.1699
SwinTSmall0.8341 ± 0.12270.9084 ± 0.07780.7817 ± 0.1671
SwinTBase0.829 ± 0.12540.8887 ± 0.08480.7874 ± 0.1701
SwinTLarge0.8358 ± 0.12370.9077 ± 0.08160.7849 ± 0.1667

References

  1. Bilodeau, M.F.; Esau, T.J.; MacEachern, C.B.; Farooque, A.A.; White, S.N.; Zaman, Q.U. Identifying Hair Fescue in Wild Blueberry Fields Using Drone Images for Precise Application of Granular Herbicide. Smart Agric. Technol. 2023, 3, 100127. [Google Scholar] [CrossRef]
  2. Chen, S.W.; Shivakumar, S.S.; Dcunha, S.; Das, J.; Okon, E.; Qu, C.; Taylor, C.J.; Kumar, V. Counting Apples and Oranges with Deep Learning: A Data-Driven Approach. IEEE Robot. Autom. Lett. 2017, 2, 781–788. [Google Scholar] [CrossRef]
  3. Fiona, J.R.; Anitha, J. Automated Detection of Plant Diseases and Crop Analysis in Agriculture Using Image Processing Techniques: A Survey. In Proceedings of the 2019 IEEE International Conference on Electrical, Computer and Communication Technologies (ICECCT), Coimbatore, India, 20–22 February 2019; pp. 1–5. [Google Scholar]
  4. Nuske, S.; Achar, S.; Bates, T.; Narasimhan, S.; Singh, S. Yield Estimation in Vineyards by Visual Grape Detection. In Proceedings of the 2011 IEEE/RSJ International Conference on Intelligent Robots and Systems, San Francisco, CA, USA, 25–30 September 2011; IEEE: Piscataway, NJ, USA, 2011; pp. 2352–2358. [Google Scholar]
  5. Cordier, A.; Gutierrez, P.; Plessis, V. Improving Generalization with Synthetic Training Data for Deep Learning Based Quality Inspection. arXiv 2022, arXiv:2202.12818. [Google Scholar] [CrossRef]
  6. Said, A.F.; Kashyap, V.; Choudhury, N.; Akhbari, F. A Cost-Effective, Fast, and Robust Annotation Tool. In Proceedings of the 2017 IEEE Applied Imagery Pattern Recognition Workshop (AIPR), Washington, DC, USA, 10–12 October 2017; pp. 1–6. [Google Scholar]
  7. MacEachern, C.B.; Esau, T.J.; Schumann, A.W.; Hennessy, P.J.; Zaman, Q.U. Detection of Fruit Maturity Stage and Yield Estimation in Wild Blueberry Using Deep Learning Convolutional Neural Networks. Smart Agric. Technol. 2023, 3, 100099. [Google Scholar] [CrossRef]
  8. Manfrini, L.; Pierpaoli, E.; Zibordi, M.; Morandi, B.; Muzzi, E.; Losciale, P.; Grappadelli, L.C. Monitoring Strategies for Precise Production of High Quality Fruit and Yield in Apple in Emilia-Romagna. Chem. Eng. Trans. 2015, 44, 301–306. [Google Scholar]
  9. MacEachern, C.B.; Esau, T.J.; Zaman, Q.U.; White, S.N.; Farooque, A.A. Development of a Novel Precision Applicator for Spot Treatment of Granular Agrochemical in Wild Blueberry. Sci. Rep. 2024, 14, 13751. [Google Scholar] [CrossRef]
  10. Zaman, Q.U.; Esau, T.J.; Schumann, A.W.; Percival, D.C.; Chang, Y.K.; Read, S.M.; Farooque, A.A. Development of Prototype Automated Variable Rate Sprayer for Real-Time Spot-Application of Agrochemicals in Wild Blueberry Fields. Comput. Electr. Agric. 2011, 76, 175–182. [Google Scholar] [CrossRef]
  11. Abbey, J.; Percival, D.; Jaakola, L.; Asiedu, S.K. Efficacy, Persistence and Residue Levels of Fungicides for Botrytis Control in Wild Blueberry. Crop Prot. 2024, 179, 106633. [Google Scholar] [CrossRef]
  12. White, S.N. Evaluation of Herbicides for Hair Fescue (Festuca filiformis) Management and Potential Seedbank Reduction in Lowbush Blueberry. Weed Technol. 2019, 33, 840–846. [Google Scholar] [CrossRef]
  13. Eaton, L.J.; Nams, V.O. Second Cropping of Wild Blueberries—Effects of Management Practices. Can. J. Plant Sci. 2006, 86, 1189–1195. [Google Scholar] [CrossRef]
  14. Burgess, P. Guide to Weed, Insect and Disease Management in Wild Blueberry; Perennia: Kentville, NS, Canada, 2018. [Google Scholar]
  15. White, S.N.; Boyd, N.S.; Van Acker, R.C. Growing Degree-Day Models for Predicting Lowbush Blueberry (Vaccinium angustifolium Ait.) Ramet Emergence, Tip Dieback, and Flowering in Nova Scotia, Canada. HortScience 2012, 47, 1014–1021. [Google Scholar] [CrossRef]
  16. Perennia Food and Agriculture Corporation. Nova Scotia Wild Blueberry Crop Development Schedule and Management Recommendations. 2024. Available online: https://www.perennia.ca/wp-content/uploads/2018/03/Nova-Scotia-Wild-Blueberry-Crop-Development-Schedule-and-Management-Recommendations-2024.pdf (accessed on 3 October 2025).
  17. Jewell, L.E.; Compton, K.; Wiseman, D. Evidence for a Genetically Distinct Population of Exobasidium sp. Causing Atypical Leaf Blight Symptoms on Lowbush Blueberry (Vaccinium angustifolium Ait.) in Newfoundland and Labrador, Canada. Can. J. Plant Pathol. 2021, 43, 897–904. [Google Scholar] [CrossRef]
  18. Percival, D.C.; Dawson, J.K. Foliar Disease Impact and Possible Control Strategies in Wild Blueberry Production. In Proceedings of the ISHS Acta Horticulturae 810: IX International Vaccinium Symposium, Corvallis, OR, USA, 13–16 July 2008; pp. 345–354. [Google Scholar]
  19. Fenu, G.; Malloci, F.M. DiaMOS Plant: A Dataset for Diagnosis and Monitoring Plant Disease. Agronomy 2021, 11, 2107. [Google Scholar] [CrossRef]
  20. Lucas, G.B.; Campbell, C.L.; Lucas, L.T. Diseases Caused by Viruses. In Introduction to Plant Diseases: Identification and Management; Springer: Boston, MA, USA, 1992; pp. 291–308. ISBN 978-0-412-06961-1. [Google Scholar]
  21. Hildebrand, P.D.; Nickerson, N.L.; McRae, K.B.; Lu, X. Incidence and Impact of Red Leaf Disease Caused by Exobasidium vaccinii in Lowbush Blueberry Fields in Nova Scotia. Can. J. Plant Pathol. 2000, 22, 364–367. [Google Scholar] [CrossRef]
  22. Lyu, H.; McLean, N.; McKenzie-Gopsill, A.; White, S.N. Weed Survey of Nova Scotia Lowbush Blueberry (Vaccinium angustifolium Ait.) Fields. Int. J. Fruit Sci. 2021, 21, 359–378. [Google Scholar] [CrossRef]
  23. MacEachern, C.B.; Esau, T.J.; White, S.N.; Zaman, Q.U.; Farooque, A.A. Evaluation of Dichlobenil for Hair Fescue (Festuca filiformis Pourr.) Management in Wild Blueberry (Vaccinium angustifolium Ait.). Agron. J. 2024, 116, 590–597. [Google Scholar] [CrossRef]
  24. Esau, T.; Zaman, Q.; Groulx, D.; Farooque, A.; Schumann, A.; Chang, Y. Machine Vision Smart Sprayer for Spot-Application of Agrochemical in Wild Blueberry Fields. Precis. Agric. 2018, 19, 770–788. [Google Scholar] [CrossRef]
  25. Li, Y.; Kong, D.; Zhang, Y.; Tan, Y.; Chen, L. Robust Deep Alignment Network with Remote Sensing Knowledge Graph for Zero-Shot and Generalized Zero-Shot Remote Sensing Image Scene Classification. ISPRS J. Photogramm. Remote Sens. 2021, 179, 145–158. [Google Scholar] [CrossRef]
  26. Mullins, C.C.; Esau, T.J.; Zaman, Q.U.; Toombs, C.L.; Hennessy, P.J. Leveraging Zero-Shot Detection Mechanisms to Accelerate Image Annotation for Machine Learning in Wild Blueberry (Vaccinium angustifolium Ait.). Agronomy 2024, 14, 2830. [Google Scholar] [CrossRef]
  27. Cheng, T.; Song, L.; Ge, Y.; Liu, W.; Wang, X.; Shan, Y. YOLO-World: Real-Time Open-Vocabulary Object Detection. arXiv 2024, arXiv:2401.17270. [Google Scholar]
  28. Rettenberger, L.; Schilling, M.; Reischl, M. Annotation Efforts in Image Segmentation Can Be Reduced by Neural Network Bootstrapping. Curr. Dir. Biomed. Eng. 2022, 8, 329–332. [Google Scholar] [CrossRef]
  29. Sánchez, J.S.; Lisani, J.-L.; Catalán, I.A.; Álvarez-Ellacuría, A. Leveraging Bounding Box Annotations for Fish Segmentation in Underwater Images. IEEE Access 2023, 11, 125984–125994. [Google Scholar] [CrossRef]
  30. Kirillov, A.; Mintun, E.; Ravi, N.; Mao, H.; Rolland, C.; Gustafson, L.; Xiao, T.; Whitehead, S.; Berg, A.C.; Lo, W.-Y. Segment Anything. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 2–3 October 2023; pp. 4015–4026. [Google Scholar]
  31. Mazurowski, M.A.; Dong, H.; Gu, H.; Yang, J.; Konz, N.; Zhang, Y. Segment Anything Model for Medical Image Analysis: An Experimental Study. Med. Image Anal. 2023, 89, 102918. [Google Scholar] [CrossRef]
  32. Yuan, Z. Principles, Applications, and Advancements of the Segment Anything Model. Appl. Comput. Eng. 2024, 53, 73–78. [Google Scholar] [CrossRef]
  33. Ren, T.; Liu, S.; Zeng, A.; Lin, J.; Li, K.; Cao, H.; Chen, J.; Huang, X.; Chen, Y.; Yan, F.; et al. Grounded SAM: Assembling Open-World Models for Diverse Visual Tasks. arXiv 2024, arXiv:2401.14159. [Google Scholar] [CrossRef]
  34. Brust, C.-A.; Käding, C.; Denzler, J. Active Learning for Deep Object Detection. arXiv 2018, arXiv:1809.09875. [Google Scholar] [CrossRef]
  35. Schmidt, S.; Rao, Q.; Tatsch, J.; Knoll, A. Advanced Active Learning Strategies for Object Detection. In Proceedings of the 2020 IEEE Intelligent Vehicles Symposium (IV), Las Vegas, NV, USA, 19 October–13 November 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 871–876. [Google Scholar]
  36. Gurav, R.; Patel, H.; Shang, Z.; Eldawy, A.; Chen, J.; Scudiero, E.; Papalexakis, E. Can SAM Recognize Crops? Quantifying the Zero-Shot Performance of a Semantic Segmentation Foundation Model on Generating Crop-Type Maps Using Satellite Imagery for Precision Agriculture. arXiv 2023, arXiv:2311.15138. [Google Scholar]
  37. Casas, G.G.; Ismail, Z.H.; Shapiai, M.I.; Karuppiah, E.K. Automated Detection and Segmentation of Baby Kale Crowns Using Grounding DINO and SAM for Data-Scarce Agricultural Applications. Smart Agric. Technol. 2025, 11, 100903. [Google Scholar] [CrossRef]
  38. Tian, Q.; Zhang, H.; Bian, L.; Zhou, L.; Shen, Z.; Ge, Y. Field-Based Phenotyping for Poplar Seedlings Biomass Evaluation Based on Zero-Shot Segmentation with Multimodal UAV Images. Comput. Electron. Agric. 2025, 236, 110462. [Google Scholar] [CrossRef]
  39. Yang, C.; Jin, Q.; Wang, Y.; Zhou, Y.; Lan, D.; Yang, Y. EHAPZero: Ensemble Hierarchical Attribute Prompting Based Zero-Shot Learning for Pest Recognition. IEEE Internet Things J. 2024, 12, 49107–49119. [Google Scholar] [CrossRef]
  40. Esmaeilpour, S.; Liu, B.; Robertson, E.; Shu, L. Zero-Shot out-of-Distribution Detection Based on the Pre-Trained Model Clip. In Proceedings of the AAAI Conference on Artificial Intelligence, Online, 22 February–1 March 2022; Volume 36, pp. 6568–6576. [Google Scholar]
  41. Ravi, N.; Gabeur, V.; Hu, Y.-T.; Hu, R.; Ryali, C.; Ma, T.; Khedr, H.; Rädle, R.; Rolland, C.; Gustafson, L.; et al. SAM 2: Segment Anything in Images and Videos. arXiv 2024, arXiv:2408.00714. [Google Scholar]
  42. Rahman, M.A.; Wang, Y. Optimizing Intersection-Over-Union in Deep Neural Networks for Image Segmentation. In Advances in Visual Computing; Bebis, G., Boyle, R., Parvin, B., Koracin, D., Porikli, F., Skaff, S., Entezari, A., Min, J., Iwai, D., Sadagic, A., et al., Eds.; Lecture Notes in Computer Science; Springer International Publishing: Cham, Switzerland, 2016; Volume 10072, pp. 234–244. ISBN 978-3-319-50834-4. [Google Scholar]
  43. Montgomery, D.C. Design and Analysis of Experiments; John Wiley & Sons: Hoboken, NJ, USA, 2017; ISBN 978-1-119-11347-8. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.