Next Article in Journal
Erosion Assessment by a Fast and Low-Cost Procedure in a Vineyard Under Different Soil Management
Previous Article in Journal
Preliminary Study of the Genetic Response of Grapevine Buds to a Preventive Natural Polysaccharide-Based Biogel Under Simulated Late Frost Conditions
Previous Article in Special Issue
A Unified Framework for Enhanced 3D Spatial Localization of Weeds via Keypoint Detection and Depth Estimation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

BudCAM: An Edge Computing Camera System for Bud Detection in Muscadine Grapevines

1
Department of Biological Systems Engineering, University of Nebraska—Lincoln, Lincoln, NE 68583, USA
2
Panhandle Research and Extension Center, University of Nebraska—Lincoln, Scottsbluff, NE 69361, USA
3
Biological Systems Engineering, College of Agriculture and Food Sciences, Florida A&M University, Tallahassee, FL 32307, USA
4
Center for Viticulture and Small Fruit Research, Florida A&M University, Tallahassee, FL 32307, USA
*
Authors to whom correspondence should be addressed.
Agriculture 2025, 15(21), 2220; https://doi.org/10.3390/agriculture15212220
Submission received: 26 September 2025 / Revised: 21 October 2025 / Accepted: 22 October 2025 / Published: 24 October 2025

Abstract

Bud break is a critical phenological stage in muscadine grapevines, marking the start of the growing season and the increasing need for irrigation management. Real-time bud detection enables irrigation to match muscadine grape phenology, conserving water and enhancing performance. This study presents BudCAM, a low-cost, solar-powered, edge computing camera system based on Raspberry Pi 5 and integrated with a LoRa radio board, developed for real-time bud detection. Nine BudCAMs were deployed at Florida A&M University Center for Viticulture and Small Fruit Research from mid-February to mid-March, 2024, monitoring three wine cultivars (A27, noble, and Floriana) with three replicates each. Muscadine grape canopy images were captured every 20 min between 7:00 and 19:00, generating 2656 high-resolution (4656 × 3456 pixels) bud break images as a database for bud detection algorithm development. The dataset was divided into 70% training, 15% validation, and 15% test. YOLOv11 models were trained using two primary strategies: a direct single-stage detector on tiled raw images and a refined two-stage pipeline that first identifies the grapevine cordon. Extensive evaluation of multiple model configurations identified the top performers for both the single-stage (mAP@0.5 = 86.0%) and two-stage (mAP@0.5 = 85.0%) approaches. Further analysis revealed that preserving image scale via tiling was superior to alternative inference strategies like resizing or slicing. Field evaluations conducted during the 2025 growing season demonstrated the system’s effectiveness, with the two-stage model exhibiting superior robustness against environmental interference, particularly lens fogging. A time-series filter smooths the raw daily counts to reveal clear phenological trends for visualization. In its final deployment, the autonomous BudCAM system captures an image, performs on-device inference, and transmits the bud count in under three minutes, demonstrating a complete, field-ready solution for precision vineyard management.

1. Introduction

Bud break represents a critical phenological stage in grapevines, marking the beginning of each new growing season and influencing both vegetative development and fruit set. In muscadine grape production, irrigation is particularly important during the first two years while vines are establishing, and after establishment, water requirements are greatest from bud break through flowering [1]. Adequate soil water availability is essential during bud break and early shoot growth, as water stress at this stage can restrict vegetative expansion, resulting in smaller canopies and reduced grape yield and quality [1]. Bud break serves not only as a vital phenological indicator but also as a practical signal for initiating irrigation in grapevine management.
Bud break monitoring in commercial vineyards largely depends on manual visual inspection. Vineyard growers typically record bud break timing, extent, and progression by scouting rows and visually counting buds, which is time consuming, labor intensive, and prone to subjectivity. The reliance on manual methods restricts monitoring efforts to small sample areas, limiting decision making and the ability to optimize vineyard management at scale. Furthermore, bud break is highly sensitive to environmental conditions, particularly spring temperature [2], making timely and accurate monitoring even more essential.
To address these limitations, recent advances in computer vision and deep learning have provided powerful tools for automating agricultural tasks. Deep learning models, which use large numbers of parameters to recognize complex patterns, have become widely applied in computer vision. One key task is object detection, which identifies and labels objects within images by drawing bounding boxes around them [3]. This approach can be adapted for agriculture, where cameras capture field images and the system detects buds among grapevines [4,5,6], supporting more efficient field monitoring.
Grimm et al. [7] proposed a Convolutional Neural Network (CNN)-based semantic segmentation approach that localizes the grapevine organs at the pixel scale and then counts them via connected component labeling. Similarly, Rudolph et al. [8] used a CNN-based segmentation model to identify inflorescence pixels and then applied a circular Hough Transform to detect and count individual flowers. Marset et al. [9] proposed a Fully Convolutional Network (FCN)-MobileNet that directly segments grapevine buds at the pixel level and benchmarked it against a strong sliding-window patch-classifier baseline [10], reporting better segmentation, correspondence, and localization. Xia et al. [6] introduced an attention-guided data enrichment network for fine-grained classification of apple bud types, using a Split-Attention Residual Network (ResNeSt-50) [11] backbone with attention-guided CutMix and erasing modules to focus learning on salient bud regions.
Beyond pixel-wise segmentation, single-stage detectors such as the You Only Look Once (YOLO) family are popular in agriculture vision due to their excellent speed–performance tradeoff and suitability for edge devices. Common strategies for low-power deployment include a lightweight backbone (e.g., GhostNet [12]), depthwise separable convolutions [13], attention modules such as the Bottleneck Attention Module (BAM) [14], enhanced multi-scale fusion, and modern Intersection over Union (IoU) losses. Li et al. [4] proposed a lightweight YOLOv4 for tea bud detection by replacing the backbone with the GhostNet and introducing depthwise separable convolutions [13] to reduce parameters and floating-point operations per second (FLOPS), which stabilized bounding box regression of small buds in clutter with Scaled Intersection over Union (SIoU) [15]. Gui et al. [5] compressed YOLOv5 using GhostConv with BAM and weighted feature fusion in the neck, retaining Complete Intersection over Union (CIoU) [16] for regression. Chen et al. [17] introduced TSP-YOLO for the monitoring of cabbage seedling emergence, modifying YOLOv8-n by replacing C2f with Swin-conv [18] for stronger multi-scale features and adding ParNet attention [19] to suppress background clutter. Pawikhum et al. [20] used YOLOv8-n for apple bud detection and aggregated detections over time downstream to avoid double counts and omissions. Taken together, these studies illustrate a consistent recipe for small-object detection on resource-constrained platforms.
Within this design space, YOLOv11 [21] offered architectural updates targeted at small-object detection under field conditions while maintaining low latency for edge deployment. The C3k2 block replaced C2f in the backbone and neck, improving computational efficiency, and the addition of a C2PSA spatial-attention block enhances the model’s ability to focus on salient regions, which is critical for detecting small buds against complex backgrounds. Reported benchmarks indicate YOLOv11 attains comparable or improved accuracy with fewer parameters than prior YOLO versions, achieving similar mean Average Precision (mAP@0.5) with 22% fewer parameters than YOLOv8. These properties align with the needs of a low-power, in-field bud-monitoring system.
This study develops a custom-built bud detection camera (BudCAM), an integrated, low-cost, solar-powered edge computing system for automated bud break monitoring. The objectives of this study are to: (1) deploy the BudCAM devices for in-field data acquisition and on-device processing; (2) develop and validate a YOLOv11-based detection framework using a bud break image database constructed in 2024 and 2025; and (3) establish a post-processing pipeline that converts raw detections into smoothed time-series data for phenological analysis. Collectively, these contributions advance automated, edge-based monitoring of grapevine bud development.

2. Materials and Methods

This section begins with an overview of the BudCAM hardware, a description of field sites, and the data collection strategy, which together form the foundation for deploying bud break detection algorithms on the custom-built camera in field applications. The BudCAM units captured images at regular intervals during the bud break period, and these images were annotated to create training datasets for YOLOv11 detection models. Different image processing strategies, model configurations, and data augmentation methods were evaluated to determine optimal performance. Model accuracy was assessed through multiple testing strategies and validated under real-world field conditions.

2.1. Hardware Development

BudCAM is mounted one meter above the grapevine to capture images of the canopy (Figure 1). The grapevine cordon is positioned approximately 1.5 m above the ground, resulting in a total camera height of about 2.5 m from the ground.
The BudCAM system integrates multiple hardware components into a low-cost, solar-powered, edge computing platform designed for field deployment in vineyards. A Raspberry Pi 5 (Raspberry Pi Foundation, Cambridge, UK), equipped with a 512 GB SD card, serves as the central processing unit, managing image capture, local execution of the trained deep learning model, and data transmission. The ArduCAM 16 MP (Raspberry Pi Foundation, Cambridge, UK) autofocus camera captures high-resolution (4656 × 3456 pixels) grapevine canopy images, ensuring sharp focus for bud detection and clear visualization. BudCAM is powered by two 3.7 V, 11,000 mAh lithium batteries connected in parallel to meet the demand of the Raspberry Pi 5 during inference. This configuration is necessary because a single battery cannot supply sufficient current when BudCAM executes the object detection algorithm. The two batteries are charged by a 6 V, 20 W solar panel via a voltage regulator that conditions the panel output and regulates constant current/constant voltage (CC/CV) charging. Power delivery to the Raspberry Pi 5 is managed by a PiJuice HAT (Raspberry Pi Foundation, Cambridge, UK) power management board, which features a Real-Time Clock (RTC) to automatically wake the system at 30 min sampling intervals during daylight hours (07:00–19:00), thereby optimizing energy use. BudCAM follows a three-step processing framework (Figure 2): (1) image acquisition and on-device processing by the Raspberry Pi 5, (2) the transmission of processed bud counts via a Maduino LoRa radio board (Makerfabs, Shenzhen, China) (915 MHz) to a custom-built Raspberry Pi 4 LoRa gateway, and (3) cloud-based visualization through a dedicated web interface (https://phrec-irrigation.com). Due to LoRa’s limited bandwidth, only per-image bud count is transmitted, while raw images remain stored locally on the SD card for offline analysis. The gateway, located within LoRa range at a powered site (e.g., a laboratory), pushes the data via the internet to the cloud server for visualization, enabling real-time vineyard monitoring through the web platform.

2.2. Experiment Site and Data Collection

This experiment was conducted at the Center for Viticulture and Small Fruit Research at Florida A&M University during the 2024 and 2025 muscadine grape growing seasons. Nine muscadine grapevines representing three cultivars (Floriana, Noble, and A27) were monitored, with three vines randomly selected as replicates for each cultivar. Each vine was equipped with a dedicated BudCAM unit. During the 2024 growing season, nine BudCAMs operated continuously from March to August (Table 1), capturing images every 20 min between 07:00 and 19:00 daily. Images were collected throughout the growing season, but only those captured during each vine’s specific bud break period were selected for algorithm development. These bud break periods varied by grapevine, ranging from late March to mid-April, as detailed in Table 1. The selected images were manually annotated for bud detection using the Roboflow (version 1.1.65) platform [22]. In this study, distinguishing between buds and shoots was particularly difficult from top-view images, where the structures often overlap and appear visually similar. To maintain consistency, we combined both categories under a single label, “bud,” during dataset annotation. This simplified the process while still capturing the key developmental stages required for monitoring grapevine growth. In 2025, the nine BudCAMs were deployed to test the bud detection algorithm developed from 2024 data. During the 2025 field testing and validation period (5 March to 20 April), the systems automatically processed images using the trained model every 30 min between 7:00 and 19:00 daily.

2.3. Software Development for BudCAM

2.3.1. Detection Algorithm

The detection framework was implemented using YOLOv11 [21], developed by Ultralytics in PyTorch and publicly available. YOLOv11 comprises three main components: the backbone, the neck, and the head. The backbone extracts multi-scale features, which are then aggregated and refined by the neck. The head uses these refined features to predict bounding box coordinates, class probabilities, and confidence scores for bud detection [21].
Two approaches were employed for both training and inference: (1) a Single-Stage Bud Detector, in which a YOLO model was trained directly to detect buds from images, outputting bounding boxes and confidence scores for each detected bud (Figure 3a); and (2) a Two-Stage Pipeline, in which a grapevine cordon detection model first identifies the main grapevine structure. When a cordon prediction exceeds a confidence threshold, the corresponding region is cropped from the original image and passed to a second YOLO model trained specifically for bud detection. Images without a confident cordon detection are skipped to avoid false detections from non-vine areas (Figure 3b).
Both detection approaches process images using a tiling strategy rather than directly analyzing full-resolution images, with details provided in Section 2.3.2. After detecting buds in each tile, each bud detection with its bounding box is mapped back to the original-image coordinates for evaluation and bud counting.
All experiments followed the default YOLOv11 training and evaluation protocol with standard objective functions. Model performance was assessed using mean Average Precision at an Intersection over Union (IoU) threshold of 0.5 (mAP@0.5), a widely adopted metric for object detection. Model development and training utilized the Ultralytics YOLOv11 framework (version 8.3.158) in Python 3.12.8 with PyTorch 2.5.1.

2.3.2. Dataset Preparation

A total of 2656 high-resolution (4656 × 3456 pixels) raw images were collected in 2024 as a bud break image database and split into 70% for training, 15% for validation, and 15% for test subsets. The distribution of raw images per vine across these three subsets is provided in Table 2, which ensures balanced representation of cultivars for model development and evaluation. From these subsets, two derived datasets were prepared using different processing strategies: the Raw Grid (RG) dataset and the Cordon-Cropped (CC) dataset. The RG dataset feeds the single-stage pipeline, whereas the CC dataset is used in the two-stage pipeline as input to the second-stage bud detector. CC images were generated by applying a separately trained cordon detector to crop the cordon region, which was then used for bud detection.
The first strategy tiled each raw image into a 6 × 4 grid (776 × 864 pixels per tile), forming the RG dataset (Table 3). The CC dataset was generated in two steps. First, a cordon detector was trained using a YOLOv11-medium model on 1500 annotated cordon images resized to 1280 × 1280 pixels, with augmentation including flips and rotations up to 180°. The model was fine-tuned from COCO-pretrained weights for 500 epochs with a batch size of 32. It achieved an mAP@0.5 of 0.99 and an mAP@0.95 of 0.86, ensuring reliable extraction of the cordon region. Using this detector, the grapevine cordon was cropped from each image and then tiled into 640 × 640 patches. These patches formed the CC dataset, which was specifically used to train the bud detector. The relationship between the two dataset preparation strategies and their role in the pipelines is summarized in Table 3.
Both of the RG and CC datasets consist of training, validation, and test subsets derived, respectively, from the same divisions of the raw images. Image tiling for both datasets was performed after annotation, and annotations were exported in YOLO format [23]. To reduce false positives and improve generalization, background-only tiles were retained so that approximately 80% of the training tiles contained visible buds. The number of samples in each subset for the RG and CC datasets is summarized in Table 3. For CC dataset generation, the cordon detector applied a confidence threshold of 0.712 (selected by maximizing F1 on the validation set). The highest-scoring cordon above the threshold was cropped, whereas images without a cordon above the threshold were excluded from the CC dataset.

2.3.3. Model Training Configuration

Several training configurations were evaluated to assess model performance, including the application of transfer learning, augmentation hyperparameter tuning, and model size variation. All configurations were trained on two NVIDIA Tesla V100 GPUs (32 GB each) at the Holland Computing Center (HCC) at the University of Nebraska–Lincoln.
Transfer learning was implemented by initializing models with weights pretrained on the COCO dataset [24], allowing them to converge faster and leverage knowledge learned from the source task and be fine-tuned for the target task. Data augmentation was performed on the fly during training, ensuring that augmented samples differed across epochs to enhance dataset diversity and model generalization. Two augmentation configurations were evaluated, with parameters randomly sampled from uniform distributions, and the details are provided in Table 4.
Configuration 1 followed the setup by Marset et al. [9], representing a commonly used augmentation scheme in object detection studies. Each raw sample was augmented by (1) rotation from 0° to 45°, (2) horizontal and vertical translation up to 0.4, (3) scaling between 0.7 and 1.3, (4) shearing up to 0.1, and (5) horizontal and vertical flips. Configuration 2 was derived from Configuration 1 to better reflect the small bud size in the images and BudCAM’s deployment, which captures a top-down view of the cordon and buds. A mild perspective transformation was added to account for foreshortening, while translation, scaling, and shear ranges were reduced to avoid excessive distortion of small buds. The rotation range was narrowed from 45° to 10° to prevent angular distortion and bounding box misalignment. These modifications reflect a trade-off between robustness to spatial variation and preserving bud visibility, critical for tiny-object detection. To investigate the effect of model capacity, the large model variant was also tested alongside the medium model.
Eight training configurations (Models 1–8; Table 5) were defined by combining augmentation setup, transfer learning, and model size. Each configuration was applied to both the RG and CC datasets, producing 16 trained models in total. For the two-stage pipeline, the CC-trained models served as the bud detector, while the cordon detector was trained separately as a preprocessing step.
All input samples were resized to 640 × 640 pixels using bilinear interpolation [25] to match the input size requirement of YOLOv11. Training was conducted with a batch size of 128 using the stochastic gradient descent (SGD) optimizer, an initial learning rate of 0.01 scheduled by cosine annealing, and a maximum of 1000 epochs. Early stopping was applied, with training terminated if the validation metric failed to improve for 20 consecutive epochs. To assess robustness, each configuration was repeated five times with different random seeds. Model performance was evaluated using mean Average Precision (mAP) at an Intersection over Union (IoU) threshold of 0.5, a standard metric in object detection that measures the agreement between predicted bounding boxes and ground-truth annotations.

2.3.4. Inference Strategies on the 2024 Dataset

Model performance was first evaluated using the baseline test set, where predictions were generated directly from a prepared test set without additional processing. To more thoroughly investigate the performance under different conditions, two additional inference strategies were applied. In both strategies, the grapevine cordon region was first extracted from the raw test images using the pretrained cordon detector.
The first strategy, referred to as the Resized Cordon Crop (Resize-CC) approach, resized each cropped cordon to the model’s 640 × 640 input size. This strategy evaluates model performance under significant downscaling, where small objects such as buds risk losing critical visual detail.
The baseline tiling strategy, while effective, may split buds located at tile boundaries, resulting in missed detections. To address this limitation, the second strategy, known as the Sliced Cordon Crop (Sliced-CC) evaluation, was introduced. This approach applied a sliding-window technique to divide each cordon crop into 640 × 640 slices with slice-overlap ratios of r = 0.2, 0.5, and 0.75. Each slice was processed individually by the model, and the resulting detections were remapped to the original image coordinates and merged using non-maximum suppression (NMS).

2.4. Setup of Real-World Testing

The trained models were deployed on nine BudCAMs during the 2025 muscadine grapevine growing season at Florida A&M University Center for Viticulture and Small Fruit Research to evaluate near real-time monitoring and data acquisition. Two types of visual evaluation were conducted: offline evaluations provided qualitative detection results across models, while online evaluations demonstrated real-world performance under field conditions. The offline evaluation used all images collected between 5 March and 20 April 2025, across nine posts, and was conducted using the optimal single-stage model (the one with the highest mAP@0.5 among Models 1–8) and the optimal two-stage model (the one with the highest mAP@0.5 among Models 1–8). The online evaluation continuously reported near real-time detection results from YOLOv11-Medium (Model 7 of single-stage) deployed on BudCAM. Processed outputs were displayed on a custom-designed website (https://phrec-irrigation.com), demonstrating the edge computing capability of the system. This field evaluation captured dynamic conditions such as changing sunlight, camera angles, and bud growth stages, providing a realistic assessment of model performance in practice.

2.5. Time-Series Post-Processing

Bud counts from the nine posts were ordered by timestamp to generate a raw time series for each vine. The raw data, however, contains spurious fluctuations due to transient imaging artifacts, such as lens fogging, glare, and exposure flicker. To reduce this noise and produce a continuous representation of the biological trend, an adaptive Exponentially Weighted Moving Average (adaptive-EWMA) smoother [26] was applied to down-weight outliers and reduce misleading spikes in the plots. Let x t denote the observed bud count and m t 1 the previous smoothed value for t = 1 , 2 , 3 , The adaptive-EWMA updates as:
m t = x t , if t = 0 m t 1 + α · ψ ( r t ) , else
where ψ ( r t ) is the Huber clipping function,
ψ ( r t ) = r t , if | r t | δ δ · s i g n ( r t ) , if | r t | > δ
where r t = x t m t 1 is the residual, which represents how much the current observation x t deviates from the expected (smoothed) value m t 1 . The Huber function treats small residuals proportionally but clips large residuals. This prevents extreme observations from disproportionately influencing the smoothed series. In this study, δ was defined as twice the robust standard deviation:
δ = 2 · σ 2 · 1.4826 · M A D
where MAD is the median absolute deviation. The constant 1.4826 is the standard consistency factor that makes MAD a robust estimator of σ under Gaussian distributions [27,28,29]. The smoothing parameter α was determined by the effective window size N e f f [30]:
α = 2 N e f f + 1
Since bud count trends were expected to vary on a daily scale, and with 22 images captured per day, N e f f was set to 22, yielding α = 0.087 . These settings were used to generate the smoothed time-series curves.

3. Results and Discussion

This section presents the evaluation of the trained models, organized into three parts. The baseline performance of all 16 model configurations, including eight trained on the RG dataset and eight trained on the CC dataset, was first compared to examine the effects of model size, fine-tuning, and data augmentation, which led to the selection of the two best models (Table 4 and Table 5). Secondly, these top models were evaluated with alternative inference strategies (Resized-CC and Sliced-CC) to determine the most effective processing approach. Finally, the performance of the selected models was illustrated with qualitative examples from the 2024 dataset, capturing both early and full bud break to confirm model robustness, compared with related work, and evaluated through a quantitative time-series analysis of 2025 field images.

3.1. Model Performance and Inference Strategy Evaluation

The comparison between medium-size models (Model 1 and Model 2) and large-size models (Model 3 and Model 4) showed only marginal differences across both datasets (Table 6). In the RG dataset, the fine-tuned medium model (Model 2, mAP@0.5 = 80.3%) outperformed the fine-tuned large model (Model 4, mAP@0.5 = 78.5%) by 1.8%, while in the CC dataset, medium and large models achieved nearly identical accuracy (79.1–79.2% without fine-tuning and 78.3–78.7% with fine-tuning). These findings indicate that increasing model size did not lead to meaningful improvements in bud detection accuracy, and the medium-size models are preferable for deployment, where comparable detection accuracy can be achieved with faster inference speed on edge devices like Raspberry Pi 5.
The effect of fine-tuning on model performance was inconsistent (Table 6). For example, in the RG dataset, fine-tuning improved accuracy in the medium model, with Model 2 (80.3%) outperforming Model 1 (75.9%) by 4.4%. However, in augmented models the effect was minimal, with Model 6 (85.3%) only slightly higher than Model 5 (85.1%). Similarly, in the CC dataset, Model 8 (85.0%) performed nearly the same as Model 7 (84.7%). Overall, fine-tuning reduced training duration due to earlier convergence but did not consistently enhance model accuracy.
Data augmentation proved to be an effective training factor (Table 6). In the RG dataset, for instance, the fine-tuned model with augmentation (Model 6, mAP@0.5 = 85.3%) significantly outperformed its counterpart without augmentation (Model 2, mAP@0.5 = 80.3%), by a 5-point mAP@0.5 improvement. A similar gain was also observed in the CC dataset, where the fine-tuned model with augmentation (Model 8, mAP@0.5 = 85.0%) exceeded the non-augmented fine-tuned model (Model 2, mAP@0.5 = 78.3%) by 6.7%. Similar improvements were observed across all comparisons, confirming the benefits of augmentation for enhancing model accuracy.
Based on these comparisons, the two best-performing models—the RG-trained Model 7 and CC-trained Model 8—were selected for evaluation using alternative inference strategies, including Resized-CC and Sliced-CC applied to the original 4K-resolution images, which is described in Section 2.4. In this extended evaluation, both models under Resized-CC and Sliced-CC settings showed reduced performance compared to the baseline (Table 7). This performance gap can be explained by differences in how the datasets were prepared. The training data for Models 7 (RG) and 8 (CC) were derived from grids or patches cropped from 4K-resolution images, which preserved sufficient detail for detecting small buds. By contrast, the Resized-CC dataset was produced directly by scaling entire 4K images down to 640 × 640 pixels. This downscaling reduced the bud size for reliable detection, thereby substantially degrading performance. Consequently, Model 7 dropped from 86.0% mAP@0.5 (baseline) to 58.2% under Resized-CC, while Model 8 decreased from 85.0% to 52.6% (Table 7).
In contrast to the Resized-CC results, the Sliced-CC evaluation produced reduced performance even though this strategy was expected to improve detection by leveraging overlapping regions. A likely explanation is an increase in false positives. Specifically, each model performed 24 predictions per image when evaluated on the RG test set, corresponding to the 6 × 4 grid of tiles. With slice prediction and nonzero overlap, even more patches were generated per image, leading to redundant predictions and thus more false positives. To further investigate this trend, two higher overlap ratios (0.5 and 0.75) were tested. As shown in Table 8, detection accuracy consistently declined as the overlap ratio increased. For Model 7 (RG), mAP@0.5 decreased from 74.8% at r = 0.2 to 72.0% at r = 0.5 and further to 66.3% at r = 0.75. A similar trend was observed for Model 8 (CC), where mAP@0.5 dropped from 80.0% at r = 0.2 to 76.4% at r = 0.5 and 69.2% at r = 0.75. These results support the hypothesis that excessive prediction redundancy from overlapping slices increases false positives and degrades overall performance.
Based on these results, the original, scale-preserving tiling procedures were confirmed as the optimal inference strategies. Therefore, the following two configurations were used for all subsequent analyses: (1) RG-trained Model 7, which applies the preprocessing step that tiles each raw image into a 6 × 4 grid; and (2) CC-trained Model 8, which operates within the two-stage pipeline where a cropped cordon region is tiled into 640 × 640 patches for bud detection. These approaches were applied consistently to analyze the test set images and the field images during the 2025 growing season.

3.2. Dataset Comparison and 2024 Processed Image Examples

Two raw images from each cultivar were selected—one captured representing the early bud break season (Figure 4a, Figure 5a, Figure 6a), and the other during the full bud break season (Figure 7a, Figure 8a, Figure 9a)—resulting in six raw images in total. Detection confidence thresholds were determined from the optimal F1 scores on the respective validation sets: 0.279 for RG-trained Model 7 and 0.369 for CC-trained Model 8. Each image was processed using RG-trained Model 7 (Figure 4b, Figure 5b, Figure 6b, Figure 7b, Figure 8b, Figure 9b) and CC-trained Model 8 (Figure 4c, Figure 5c, Figure 6c, Figure 7c, Figure 8c, Figure 9c). Both models successfully detected buds under both early- and full-season conditions. The selected examples span different times of day (morning, midday, and evening), demonstrating model robustness under varying lighting conditions.

3.3. Comparison with Related Methods

The recent studies on bud, flower, and seedling detection or segmentation are summarized in Table 9. Although datasets, viewpoints, and evaluation metrics differ substantially, segmentation studies often report F1 scores, whereas detection frameworks typically use mAP@0.5. Despite these differences among studies, the comparison offers useful insight into the relative performance of existing methods. Segmentation-based studies on buds and flowers in crops such as grapevine and apple generally achieved F1 scores between 75% and 92%, demonstrating strong classification capability when large, high-quality datasets and powerful backbones were used [6,7,8,9]. Detection-based approaches, including those for tea and grapevine buds, typically reported mAP@0.5 values ranging from 85% to 99% under controlled or semi-field conditions [4,5,17,20]. Most prior works (Table 9) appear to have been tested on servers or workstations, as the platform was not specified. Only Pawikhum et al. and our work have reported embedded/edge inference [20]. Pawikhum’s results achieved an mAP@0.5 of 59.0 on a compact edge platform, whereas BudCAM, running on a Raspberry Pi 5, reached 85–86 of mAP@0.5 across two independently trained YOLOv11-m models (RG-trained Model 7 and CC-trained Model 8). Despite using 4K nadir images collected under field conditions, BudCAM delivered accuracy comparable to or exceeding that of server-based approaches. These results demonstrate that high detection accuracy can be maintained on a low-cost, energy-efficient embedded platform, enabling real-time, scalable bud monitoring for grapevine phenotyping and broader agricultural applications.

3.4. Real-World Inference in 2025

RG-trained Model 7 and CC-trained Model 8 were further applied to perform inference on images captured between 5 March and 20 April 2025, aligning with the bud season. The raw and smoothed bud counts are shown in Figure 10 and Figure 11.
Analysis of the RG-trained results (Figure 10) shows substantial noise, including severe outliers inconsistent with the early stage of bud break. For example, vine NP6 registered an anomalous peak count of 51 on 14 March 2025, near the onset of bud break. These outliers were influential enough to bias the adaptive-EWMA smoother, pulling the smoothed value up to 49 in this case. This indicates that while the smoother attenuates minor fluctuations, it can be compromised by large, persistent false positives. Inspection of corresponding images (Figure 12) suggests that most anomalies resulted from false detections caused by lens fogging in the early morning.
In contrast, the CC-trained results (Figure 11) exhibit fewer such anomalies, and all nine plots show the expected increase in bud count over time. The improvement stems from the cordon detection stage, which filters out frames where a cordon cannot be confidently localized, reducing the likelihood that fogged lens images proceed to bud detection. The adaptive-EWMA smoother was applied to stabilize residual variability at the day scale. Nevertheless, some noise remains (e.g., vine NP2 around 10 April 2025). Review of the corresponding images showed that CC-trained Model 8 underperformed in certain conditions (Figure 13), such as relatively dark images or those affected by rainbow lens flare. In these cases, buds were under-detected, and occasionally entirely missed (e.g., the left panel of Figure 13). The failure is likely because such poor-quality images were excluded during dataset preparation, leaving the model insufficiently trained to generalize to these unseen scenarios, especially when buds appear small relative to the overall image. This underestimation contributes to sharp downward excursions especially late in the season where isolated frames drop to zero or near-zero while adjacent frames report typical counts (10–40).
Finally, to illustrate the deployed system, Figure 14 shows the dashboard view for vine NP6, plotting raw counts at 30 min intervals between 3 March and 20 April 2025. In addition, the BudCAM requires approximately three minutes to complete a full operational cycle, including waking from sleep, capturing and processing an image, transmitting the bud count, and returning to sleep.

4. Conclusions

Bud break marks the transition from dormancy to active growth, when grapevine water requirements increase sharply. Near real-time bud detection provides critical information for irrigation decision making by aligning water applications with the actual phenological stage of muscadine grape. To address this need, this work introduced BudCAM, a low-cost, solar-powered edge computing system designed to provide near real-time bud monitoring in muscadine vineyards at 30 min intervals. This study evaluated 16 YOLOv11 model configurations across two distinct approaches: a single-stage detector using a Raw Grid (RG) dataset and a two-stage pipeline using a Cordon-Cropped (CC) dataset.
The top models from each approach, RG-trained Model 7 (mAP@0.5 = 86.0%) and CC-trained Model 8 (mAP@0.5 = 85.0%), were evaluated against alternative inference strategies (Resized-CC and Sliced-CC). This analysis confirmed that the original, scale-preserving tiling methods were superior, highlighting that preserving bud scale is critical for accurate detection. In-field evaluations using data collected during the 2025 muscadine grape growing season at Florida A&M University Center for the CC-trained two-stage model showed greater robustness to environmental noise like lens fog, though its performance was reduced in challenging lighting conditions such as dark scenes or glare.
Taken together, these results indicate that reliable, on-device bud detection is now feasible at field scale. By applying time-series smoothing to filter noise from the raw daily counts, the BudCAM system enables growers to track bud break progression and align early-season irrigation with phenological demand. Future work should prioritize improving model robustness to challenging lighting conditions and enhancing generalization across different sites and cultivars. While each BudCAM currently monitors a single static vine, its lightweight, solar-powered design and short duty cycle enable integration with mobile platforms for rotational or route-based monitoring, capturing spatial and temporal variation with fewer devices. Beyond muscadine grapes, BudCAM has been successfully deployed in a bunch grape vineyard in eastern Nebraska, demonstrating cross-species adaptability. The hardware and detection pipeline are retrainable for other perennial crops with visible phenological events, including bud break, flowering, and fruit development. This versatility positions BudCAM as a low-cost, generalizable platform for automated phenological monitoring across diverse vineyard and orchard systems.

Author Contributions

Writing—review and editing, writing—original draft preparation, software, visualization, methodology, investigation, formal analysis, data curation, and validation, C.-E.C.; writing—review and editing, data curation, visualization, validation, supervision, resources, project administration, methodology, funding acquisition, formal analysis, and conceptualization, W.-Z.L.; writing—review and editing, supervision, data curation, and funding acquisition, J.C.; writing—review and editing, supervision, data curation, and funding acquisition, X.Q.; funding acquisition and supervision, V.T.; data curation, visualization, methodology, and software, Z.Y.; visualization and software, J.O. All authors have read and agreed to the published version of the manuscript.

Funding

This research is supported by the Agriculture and Food Research Initiative Competitive Grants Program, Foundational and Applied Science, project award no. 2023-69017-40560, from the U.S. Department of Agriculture’s National Institute of Food and Agriculture, and by the U.S. National Science Foundation Experiential Learning for Emerging and Novel Technologies (ExLENT) Program (award no. 2322535).

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Acknowledgments

The authors gratefully acknowledge the support of the U.S. Department of Agriculture’s National Institute of Food and Agriculture, and the U.S. National Science Foundation. We also would like to extend our gratitude to Thomas Rose and Logan Van Anne for their invaluable support in in designing the ground-based camera mounts and constructing the grapevine posts used for testing prior to delivering the cameras to Florida.

Conflicts of Interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

References

  1. Hickey, C.C.; Smith, E.D.; Cao, S.; Conner, P. Muscadine (Vitis rotundifolia Michx., syn. Muscandinia rotundifolia (Michx.) Small): The resilient, native grape of the southeastern US. Agriculture 2019, 9, 131. [Google Scholar] [CrossRef]
  2. Reta, K.; Netzer, Y.; Lazarovitch, N.; Fait, A. Canopy management practices in warm environment vineyards to improve grape yield and quality in a changing climate. A review A vademecum to vine canopy management under the challenge of global warming. Sci. Hortic. 2025, 341, 113998. [Google Scholar] [CrossRef]
  3. Zhang, H.; Li, F.; Liu, S.; Zhang, L.; Su, H.; Zhu, J.; Ni, L.M.; Shum, H.Y. Dino: Detr with improved denoising anchor boxes for end-to-end object detection. arXiv 2022, arXiv:2203.03605. [Google Scholar]
  4. Li, J.; Li, J.; Zhao, X.; Su, X.; Wu, W. Lightweight detection networks for tea bud on complex agricultural environment via improved YOLO v4. Comput. Electron. Agric. 2023, 211, 107955. [Google Scholar] [CrossRef]
  5. Gui, Z.; Chen, J.; Li, Y.; Chen, Z.; Wu, C.; Dong, C. A lightweight tea bud detection model based on Yolov5. Comput. Electron. Agric. 2023, 205, 107636. [Google Scholar] [CrossRef]
  6. Xia, X.; Chai, X.; Zhang, N.; Sun, T. Visual classification of apple bud-types via attention-guided data enrichment network. Comput. Electron. Agric. 2021, 191, 106504. [Google Scholar] [CrossRef]
  7. Grimm, J.; Herzog, K.; Rist, F.; Kicherer, A.; Toepfer, R.; Steinhage, V. An adaptable approach to automated visual detection of plant organs with applications in grapevine breeding. Biosyst. Eng. 2019, 183, 170–183. [Google Scholar] [CrossRef]
  8. Rudolph, R.; Herzog, K.; Töpfer, R.; Steinhage, V. Efficient identification, localization and quantification of grapevine inflorescences and flowers in unprepared field images using Fully Convolutional Networks. Vitis 2019, 58, 95–104. [Google Scholar]
  9. Marset, W.V.; Pérez, D.S.; Díaz, C.A.; Bromberg, F. Towards practical 2D grapevine bud detection with fully convolutional networks. Comput. Electron. Agric. 2021, 182, 105947. [Google Scholar] [CrossRef]
  10. Pérez, D.S.; Bromberg, F.; Diaz, C.A. Image classification for detection of winter grapevine buds in natural conditions using scale-invariant features transform, bag of features and support vector machines. Comput. Electron. Agric. 2017, 135, 81–95. [Google Scholar] [CrossRef]
  11. Zhang, H.; Wu, C.; Zhang, Z.; Zhu, Y.; Lin, H.; Zhang, Z.; Sun, Y.; He, T.; Mueller, J.; Manmatha, R.; et al. Resnest: Split-attention networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 2736–2746. [Google Scholar]
  12. Khanam, R.; Hussain, M. What is YOLOv5: A deep look into the internal features of the popular object detector. arXiv 2024, arXiv:2407.20892. [Google Scholar] [CrossRef]
  13. Howard, A.G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H. Mobilenets: Efficient convolutional neural networks for mobile vision applications. arXiv 2017, arXiv:1704.04861. [Google Scholar] [CrossRef]
  14. Park, J.; Woo, S.; Lee, J.Y.; Kweon, I.S. Bam: Bottleneck attention module. arXiv 2018, arXiv:1807.06514. [Google Scholar] [CrossRef]
  15. Gevorgyan, Z. SIoU loss: More powerful learning for bounding box regression. arXiv 2022, arXiv:2205.12740. [Google Scholar] [CrossRef]
  16. Zheng, Z.; Wang, P.; Ren, D.; Liu, W.; Ye, R.; Hu, Q.; Zuo, W. Enhancing geometric factors in model learning and inference for object detection and instance segmentation. IEEE Trans. Cybern. 2021, 52, 8574–8586. [Google Scholar] [CrossRef]
  17. Chen, X.; Liu, T.; Han, K.; Jin, X.; Wang, J.; Kong, X.; Yu, J. TSP-yolo-based deep learning method for monitoring cabbage seedling emergence. Eur. J. Agron. 2024, 157, 127191. [Google Scholar] [CrossRef]
  18. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 10012–10022. [Google Scholar]
  19. Goyal, A.; Bochkovskiy, A.; Deng, J.; Koltun, V. Non-deep networks. Adv. Neural Inf. Process. Syst. 2022, 35, 6789–6801. [Google Scholar]
  20. Pawikhum, K.; Yang, Y.; He, L.; Heinemann, P. Development of a Machine vision system for apple bud thinning in precision crop load management. Comput. Electron. Agric. 2025, 236, 110479. [Google Scholar] [CrossRef]
  21. Khanam, R.; Hussain, M. Yolov11: An overview of the key architectural enhancements. arXiv 2024, arXiv:2410.17725. [Google Scholar] [CrossRef]
  22. Roboflow. 2025. Available online: https://roboflow.com/ (accessed on 1 January 2025).
  23. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  24. Liang, W.Z.; Oboamah, J.; Qiao, X.; Ge, Y.; Harveson, B.; Rudnick, D.R.; Wang, J.; Yang, H.; Gradiz, A. CanopyCAM–an edge-computing sensing unit for continuous measurement of canopy cover percentage of dry edible beans. Comput. Electron. Agric. 2023, 204, 107498. [Google Scholar] [CrossRef]
  25. Press, W.H. Numerical Recipes: The Art of Scientific Computing, 3rd ed.; Cambridge University Press: Cambridge, UK, 2007. [Google Scholar]
  26. Zaman, B.; Lee, M.H.; Riaz, M.; Abujiya, M.R. An adaptive approach to EWMA dispersion chart using Huber and Tukey functions. Qual. Reliab. Eng. Int. 2019, 35, 1542–1581. [Google Scholar] [CrossRef]
  27. Huber, P.J. Robust statistics. In International Encyclopedia of Statistical Science; Springer: Berlin/Heidelberg, Germany, 2011; pp. 1248–1251. [Google Scholar]
  28. Rousseeuw, P.J.; Croux, C. Alternatives to the median absolute deviation. J. Am. Stat. Assoc. 1993, 88, 1273–1283. [Google Scholar] [CrossRef]
  29. Leys, C.; Ley, C.; Klein, O.; Bernard, P.; Licata, L. Detecting outliers: Do not use standard deviation around the mean, use absolute deviation around the median. J. Exp. Soc. Psychol. 2013, 49, 764–766. [Google Scholar] [CrossRef]
  30. Brown, R.G. Smoothing, Forecasting and Prediction of Discrete Time Series; Courier Corporation: North Chelmsford, MA, USA, 2004. [Google Scholar]
Figure 1. (Left): Illustration of BudCAM deployed in the research vineyard at Florida A&M University Center for Viticulture and Small Fruit Research. (Right): Components of BudCAM include: (1) ArduCAM 16 MP autofocus camera, (2) Maduino LoRa radio board, (3) two 3.7 V lithium batteries, (4) Raspberry pi 5 and PiJuice power management board, (5) LoRa antenna, (6) voltage regulator, and (7) solar cable.
Figure 1. (Left): Illustration of BudCAM deployed in the research vineyard at Florida A&M University Center for Viticulture and Small Fruit Research. (Right): Components of BudCAM include: (1) ArduCAM 16 MP autofocus camera, (2) Maduino LoRa radio board, (3) two 3.7 V lithium batteries, (4) Raspberry pi 5 and PiJuice power management board, (5) LoRa antenna, (6) voltage regulator, and (7) solar cable.
Agriculture 15 02220 g001
Figure 2. Overview of three-phase processing framework for developing BudCAM. The workflow consists of image acquisition, on-device processing, and transmission of bud counts to a cloud platform. A rectangular callout highlights a sample region of the grapevine canopy, with an inset showing a zoomed-in view of buds within that region for enhanced visual clarity.
Figure 2. Overview of three-phase processing framework for developing BudCAM. The workflow consists of image acquisition, on-device processing, and transmission of bud counts to a cloud platform. A rectangular callout highlights a sample region of the grapevine canopy, with an inset showing a zoomed-in view of buds within that region for enhanced visual clarity.
Agriculture 15 02220 g002
Figure 3. Overview of the two detection pipelines. (a) Single-stage bud detector: the raw image is divided into a 6 × 4 grid (tiles of 776 × 864 pixels); each tile is processed by a YOLOv11 bud detector. Tile-level detections are mapped back to the full image and merged to obtain the final bud counts. (b) Two-stage pipeline: a YOLOv11 cordon detector first localizes the grapevine cordon, which is then cropped and partitioned into 640 × 640 pixels patches. These patches are processed by a second YOLOv11 model, and the results are merged. In all panels, detected buds are shown in red and the detected cordon is shown in blue.
Figure 3. Overview of the two detection pipelines. (a) Single-stage bud detector: the raw image is divided into a 6 × 4 grid (tiles of 776 × 864 pixels); each tile is processed by a YOLOv11 bud detector. Tile-level detections are mapped back to the full image and merged to obtain the final bud counts. (b) Two-stage pipeline: a YOLOv11 cordon detector first localizes the grapevine cordon, which is then cropped and partitioned into 640 × 640 pixels patches. These patches are processed by a second YOLOv11 model, and the results are merged. In all panels, detected buds are shown in red and the detected cordon is shown in blue.
Agriculture 15 02220 g003
Figure 4. Images from the A27 cultivar captured in the early bud break season, showing inference results from RG-trained Model 7 and CC-trained Model 8. (a) Raw image with ground-truth buds (pink bounding boxes) and total bud counts; the capture time is marked in the top-right corner. (b) Detection results from Model 7, with red bounding boxes indicating detected buds and associated confidence scores. (c) Detection results from Model 8, where the blue bounding box marks the cordon region identified by the cordon detector, and red bounding boxes denote detected buds with confidence scores. Bud counts in (b,c) represent the total detections per image.
Figure 4. Images from the A27 cultivar captured in the early bud break season, showing inference results from RG-trained Model 7 and CC-trained Model 8. (a) Raw image with ground-truth buds (pink bounding boxes) and total bud counts; the capture time is marked in the top-right corner. (b) Detection results from Model 7, with red bounding boxes indicating detected buds and associated confidence scores. (c) Detection results from Model 8, where the blue bounding box marks the cordon region identified by the cordon detector, and red bounding boxes denote detected buds with confidence scores. Bud counts in (b,c) represent the total detections per image.
Agriculture 15 02220 g004
Figure 5. Images from the Floriana cultivar captured in the early bud break season, showing inference results from RG-trained Model 7 and CC-trained Model 8. (a) Raw image with ground-truth buds (pink bounding boxes) and total bud counts; the capture time is marked in the top-right corner. (b) Detection results from Model 7, with red bounding boxes indicating detected buds and associated confidence scores. (c) Detection results from Model 8, where the blue bounding box marks the cordon region identified by the cordon detector, and red bounding boxes denote detected buds with confidence scores. Bud counts in (b,c) represent the total detections per image.
Figure 5. Images from the Floriana cultivar captured in the early bud break season, showing inference results from RG-trained Model 7 and CC-trained Model 8. (a) Raw image with ground-truth buds (pink bounding boxes) and total bud counts; the capture time is marked in the top-right corner. (b) Detection results from Model 7, with red bounding boxes indicating detected buds and associated confidence scores. (c) Detection results from Model 8, where the blue bounding box marks the cordon region identified by the cordon detector, and red bounding boxes denote detected buds with confidence scores. Bud counts in (b,c) represent the total detections per image.
Agriculture 15 02220 g005
Figure 6. Images from the Noble cultivar captured in the early bud break season, showing inference results from RG-trained Model 7 and CC-trained Model 8. (a) Raw image with ground-truth buds (pink bounding boxes) and total bud counts; the capture time is marked in the top-right corner. (b) Detection results from Model 7, with red bounding boxes indicating detected buds and associated confidence scores. (c) Detection results from Model 8, where the blue bounding box marks the cordon region identified by the cordon detector, and red bounding boxes denote detected buds with confidence scores. Bud counts in (b,c) represent the total detections per image.
Figure 6. Images from the Noble cultivar captured in the early bud break season, showing inference results from RG-trained Model 7 and CC-trained Model 8. (a) Raw image with ground-truth buds (pink bounding boxes) and total bud counts; the capture time is marked in the top-right corner. (b) Detection results from Model 7, with red bounding boxes indicating detected buds and associated confidence scores. (c) Detection results from Model 8, where the blue bounding box marks the cordon region identified by the cordon detector, and red bounding boxes denote detected buds with confidence scores. Bud counts in (b,c) represent the total detections per image.
Agriculture 15 02220 g006
Figure 7. Images from the A27 cultivar captured in the full bud break season, showing inference results from RG-trained Model 7 and CC-trained Model 8. (a) Raw image with ground-truth buds (pink bounding boxes) and total bud counts; the capture time is marked in the top-right corner. (b) Detection results from Model 7, with red bounding boxes indicating detected buds and associated confidence scores. (c) Detection results from Model 8, where the blue bounding box marks the cordon region identified by the cordon detector, and red bounding boxes denote detected buds with confidence scores. Bud counts in (b,c) represent the total detections per image.
Figure 7. Images from the A27 cultivar captured in the full bud break season, showing inference results from RG-trained Model 7 and CC-trained Model 8. (a) Raw image with ground-truth buds (pink bounding boxes) and total bud counts; the capture time is marked in the top-right corner. (b) Detection results from Model 7, with red bounding boxes indicating detected buds and associated confidence scores. (c) Detection results from Model 8, where the blue bounding box marks the cordon region identified by the cordon detector, and red bounding boxes denote detected buds with confidence scores. Bud counts in (b,c) represent the total detections per image.
Agriculture 15 02220 g007
Figure 8. Images from the Floriana cultivar captured in the full bud break season, showing inference results from RG-trained Model 7 and CC-trained Model 8. (a) Raw image with ground-truth buds (pink bounding boxes) and total bud counts; the capture time is marked in the top-right corner. (b) Detection results from Model 7, with red bounding boxes indicating detected buds and associated confidence scores. (c) Detection results from Model 8, where the blue bounding box marks the cordon region identified by the cordon detector, and red bounding boxes denote detected buds with confidence scores. Bud counts in (b,c) represent the total detections per image.
Figure 8. Images from the Floriana cultivar captured in the full bud break season, showing inference results from RG-trained Model 7 and CC-trained Model 8. (a) Raw image with ground-truth buds (pink bounding boxes) and total bud counts; the capture time is marked in the top-right corner. (b) Detection results from Model 7, with red bounding boxes indicating detected buds and associated confidence scores. (c) Detection results from Model 8, where the blue bounding box marks the cordon region identified by the cordon detector, and red bounding boxes denote detected buds with confidence scores. Bud counts in (b,c) represent the total detections per image.
Agriculture 15 02220 g008
Figure 9. Images from the Noble cultivar captured in the full bud break season, showing inference results from RG-trained Model 7 and CC-trained Model 8. (a) Raw image with ground-truth buds (pink bounding boxes) and total bud counts; the capture time is marked in the top-right corner. (b) Detection results from Model 7, with red bounding boxes indicating detected buds and associated confidence scores. (c) Detection results from Model 8, where the blue bounding box marks the cordon region identified by the cordon detector, and red bounding boxes denote detected buds with confidence scores. Bud counts in (b,c) represent the total detections per image.
Figure 9. Images from the Noble cultivar captured in the full bud break season, showing inference results from RG-trained Model 7 and CC-trained Model 8. (a) Raw image with ground-truth buds (pink bounding boxes) and total bud counts; the capture time is marked in the top-right corner. (b) Detection results from Model 7, with red bounding boxes indicating detected buds and associated confidence scores. (c) Detection results from Model 8, where the blue bounding box marks the cordon region identified by the cordon detector, and red bounding boxes denote detected buds with confidence scores. Bud counts in (b,c) represent the total detections per image.
Agriculture 15 02220 g009
Figure 10. Inference results from RG-trained Model 7 applied to bud season images collected between 3 March and 20 April 2025, across nine BudCAM units deployed in the vineyard. Black dots represent the raw bud counts and red dots represent the smoothed values obtained by the adaptive Exponentially Weighted Moving Average (adaptive-EWMA) filter.
Figure 10. Inference results from RG-trained Model 7 applied to bud season images collected between 3 March and 20 April 2025, across nine BudCAM units deployed in the vineyard. Black dots represent the raw bud counts and red dots represent the smoothed values obtained by the adaptive Exponentially Weighted Moving Average (adaptive-EWMA) filter.
Agriculture 15 02220 g010
Figure 11. Inference results from CC-trained Model 8 evaluated on the same vines and time period as those shown in Figure 10. Labeling and color conventions are identical.
Figure 11. Inference results from CC-trained Model 8 evaluated on the same vines and time period as those shown in Figure 10. Labeling and color conventions are identical.
Agriculture 15 02220 g011
Figure 12. Example images of three cultivars captured with a fogged lens and processed by RG-trained Model 7. Detected buds are highlighted with red bounding boxes. The capture time is shown in the top-left corner of each image.
Figure 12. Example images of three cultivars captured with a fogged lens and processed by RG-trained Model 7. Detected buds are highlighted with red bounding boxes. The capture time is shown in the top-left corner of each image.
Agriculture 15 02220 g012
Figure 13. Example of under-detected images processed by CC-trained Model 8. Detected buds are highlighted with red bounding boxes, the non-detected buds are highlighted with orange bounding boxes, and the blue bounding box marks the cordon region identified by the cordon detector.
Figure 13. Example of under-detected images processed by CC-trained Model 8. Detected buds are highlighted with red bounding boxes, the non-detected buds are highlighted with orange bounding boxes, and the blue bounding box marks the cordon region identified by the cordon detector.
Agriculture 15 02220 g013
Figure 14. Example dashboard view (NP6) showing raw counts at 30 min intervals (3 March 2025). Results were produced by RG-trained Model 7. Available at https://phrec-irrigation.com/#/f/124/sensors/410 (Access on 31 August 2025).
Figure 14. Example dashboard view (NP6) showing raw counts at 30 min intervals (3 March 2025). Results were produced by RG-trained Model 7. Available at https://phrec-irrigation.com/#/f/124/sensors/410 (Access on 31 August 2025).
Agriculture 15 02220 g014
Table 1. BudCAM node information including cultivar, harvest date, and bud break start and end dates.
Table 1. BudCAM node information including cultivar, harvest date, and bud break start and end dates.
Vine IDCultivarHarvest DateStart DateEnd Date
FP2Floriana15 August 202424 March 202415 April 2024
FP4Floriana15 August 202425 March 20248 April 2024
FP6Floriana15 August 202426 March 20249 April 2024
NP2Noble15 August 202426 March 20249 April 2024
NP4Noble15 August 202426 March 202413 April 2024
NP6Noble15 August 202424 March 202411 April 2024
A27P15A2720 August 20242 April 202420 April 2024
A27P16A2720 August 202430 March 202415 April 2024
A27P17A2720 August 20241 April 202420 April 2024
Table 2. Distribution of raw images per vine from three muscadine grape cultivars used for training, validation, and testing in model development. A total of 2656 high-resolution images were collected from nine vines and divided into 70% training, 15% validation, and 15% test subsets.
Table 2. Distribution of raw images per vine from three muscadine grape cultivars used for training, validation, and testing in model development. A total of 2656 high-resolution images were collected from nine vines and divided into 70% training, 15% validation, and 15% test subsets.
Vine IDCultivarTrainingValidationTestTotal
FP2Floriana1685243263
FP4Floriana1332430187
FP6Floriana2335147331
NP2Noble1773342252
NP4Noble2998065444
NP6Noble951917131
A27P15A273216174456
A27P16A2791910110
A27P17A273436970482
Floriana total534127120781
Noble total571132124827
A27 total7551391541048
Table 3. Summary of the Raw Grid (RG) and Cordon-Cropped (CC) datasets, including processing strategies, pipeline usage, and the number of samples in each subset after filtering. Both datasets were derived from 2656 high-resolution images collected from nine muscadine vines across three cultivars. The RG dataset was generated by tiling each raw image into 6 × 4 grids (24 tiles per image), while the CC dataset was generated by cropping the detected cordon region into 640 × 640-pixel patches for the two-stage pipeline.
Table 3. Summary of the Raw Grid (RG) and Cordon-Cropped (CC) datasets, including processing strategies, pipeline usage, and the number of samples in each subset after filtering. Both datasets were derived from 2656 high-resolution images collected from nine muscadine vines across three cultivars. The RG dataset was generated by tiling each raw image into 6 × 4 grids (24 tiles per image), while the CC dataset was generated by cropping the detected cordon region into 640 × 640-pixel patches for the two-stage pipeline.
DatasetProcessing StrategyPipeline UsageTrainingValidationTestTotal
RGRaw image tiled into 6 × 4 grid (776 × 864 px per tile)Single-stage detector15,8723387330822,567
CCCordon cropped and tiled into 640 × 640 patchesTwo-stage pipeline19,6664147411927,956
Table 4. Parameter setting for two data augmentation configurations.
Table 4. Parameter setting for two data augmentation configurations.
Config. 1RotationTranslateScaleShearPerspectiveFlip
Config. 145°40%30%10%0Yes
Config. 210°10%05%0.001Yes
1 Configuration.
Table 5. Overview of the eight training configurations with variations in model size, fine-tuning, and data augmentation strategies.
Table 5. Overview of the eight training configurations with variations in model size, fine-tuning, and data augmentation strategies.
Model IDModel SizeFine-TunedData Augmentation
Model 1MediumNoNone
Model 2MediumYesNone
Model 3LargeNoNone
Model 4LargeYesNone
Model 5MediumNoConfig 1
Model 6MediumYesConfig 1
Model 7MediumNoConfig 2
Model 8MediumYesConfig 2
Table 6. Performance and final training epochs for all model configurations trained on the Raw Grid (RG) and Cordon-Cropped (CC) datasets. Results were evaluated on test subsets from 2656 images collected from nine muscadine grapevines representing three cultivars with three replicates each.
Table 6. Performance and final training epochs for all model configurations trained on the Raw Grid (RG) and Cordon-Cropped (CC) datasets. Results were evaluated on test subsets from 2656 images collected from nine muscadine grapevines representing three cultivars with three replicates each.
Model IDFinal Training EpochTest mAP@0.5 (%)
Trained on Raw Grid (RG) Dataset
Model 170 ± 2575.9 ± 2.4
Model 279 ± 3480.3 ± 1.1
Model 377 ± 2576.0 ± 1.5
Model 469 ± 2478.5 ± 1.6
Model 5315 ± 6985.1 ± 0.4
Model 6229 ± 3085.3 ± 0.2
Model 7441 ± 7086.0 ± 0.1
Model 8238 ± 4385.7 ± 0.1
Trained on Cordon-Cropped (CC) Dataset
Model 177 ± 479.1 ± 0.3
Model 272 ± 1078.3 ± 0.3
Model 376 ± 479.2 ± 0.1
Model 471 ± 978.7 ± 0.2
Model 5512 ± 6084.4 ± 0.1
Model 6297 ± 10384.4 ± 0.4
Model 7435 ± 2784.7 ± 0.1
Model 8278 ± 4585.0 ± 0.1
All numerical values are reported as mean ± standard deviation over five runs. Bold value indicates the best-performing model in each dataset group.
Table 7. Performance comparison of the top models under different inference strategies.
Table 7. Performance comparison of the top models under different inference strategies.
ModelBaseline aResized-CCSliced-CC b
Model 7 (RG)86.0 ± 0.158.2 ± 1.374.8 ± 0.7
Model 8 (CC)85.0 ± 0.152.6 ± 1.080.0 ± 0.2
All scores are mAP@0.5 (%) reported as mean ± standard deviation. Bold values indicate the best performance for each model. a Performance on the model’s corresponding standard test set (RG or CC), which uses a tiling approach. b Sliding-window inference with a slice-overlap ratio of 0.2.
Table 8. Comparison of mAP@0.5 scores (%, mean ± SD) under three slice-overlap ratios for the top-performing models.
Table 8. Comparison of mAP@0.5 scores (%, mean ± SD) under three slice-overlap ratios for the top-performing models.
ModelSlice-Overlap Ratio
0.2 0.5 0.75
Model 7 (RG)74.8 ± 0.772.0 ± 0.766.3 ± 0.6
Model 8 (CC)80.0 ± 0.276.4 ± 0.469.2 ± 0.5
All scores are mAP@0.5 (%) reported as mean ± standard deviation. Bold values indicate the best performance for each model.
Table 9. Comparison of methods related to bud/seedling detection or segmentation.
Table 9. Comparison of methods related to bud/seedling detection or segmentation.
MethodModelDatasetView/Camera AnglePlatformPerformance
BudCAM (this work)YOLOv11-m (RG, M7) a2656 images; 4656 × 3456 pxNadir, 1–3 m above vinesRaspberry Pi 586.0 (mAP@0.5)
YOLOv11-m (CC, M8) b85.0 (mAP@0.5)
Grimm et al. [7]UNet-style (VGG16 encoder)542 images; up to 5472 × 3648 pxSide view, within ∼2 m79.7 (F1)
Rudolph et al. [8]FCN108 images; 5462 × 3648 pxSide view, within ∼1 m75.2 (F1)
Marset et al. [9]FCN–MobileNet698 images; 1024 × 1024 px88.6 (F1)
Xia et al. [6]ResNeSt5031,158 images; up to 5184 × 3156 px0.4–0.8 m92.4 (F1)
Li et al. [4]YOLOv47723 images; 1920 × 1080 px 45 90 85.2 (mAP@0.5)
Gui et al. [5]YOLOv5-l1000 images; 1600 × 1200 px92.7 (mAP@0.5)
Chen et al. [17]YOLOv8-n4058 images; 5472 × 3648 pxUAV top view, ∼3 m AGL99.4 (mAP@0.5)
Pawikhum et al. [20]YOLOv8-n1500 imagesMulti-angles, ∼0.5 mEdge (embedded)59.0 (mAP@0.5)
All scores are as reported by the respective papers; tasks and metrics differ (detection vs. segmentation; F1 vs. mAP) and are not strictly comparable. a RG-trained Model 7. b CC-trained Model 8.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chiang, C.-E.; Liang, W.-Z.; Chen, J.; Qiao, X.; Tsolova, V.; Yang, Z.; Oboamah, J. BudCAM: An Edge Computing Camera System for Bud Detection in Muscadine Grapevines. Agriculture 2025, 15, 2220. https://doi.org/10.3390/agriculture15212220

AMA Style

Chiang C-E, Liang W-Z, Chen J, Qiao X, Tsolova V, Yang Z, Oboamah J. BudCAM: An Edge Computing Camera System for Bud Detection in Muscadine Grapevines. Agriculture. 2025; 15(21):2220. https://doi.org/10.3390/agriculture15212220

Chicago/Turabian Style

Chiang, Chi-En, Wei-Zhen Liang, Jingqiu Chen, Xin Qiao, Violeta Tsolova, Zonglin Yang, and Joseph Oboamah. 2025. "BudCAM: An Edge Computing Camera System for Bud Detection in Muscadine Grapevines" Agriculture 15, no. 21: 2220. https://doi.org/10.3390/agriculture15212220

APA Style

Chiang, C.-E., Liang, W.-Z., Chen, J., Qiao, X., Tsolova, V., Yang, Z., & Oboamah, J. (2025). BudCAM: An Edge Computing Camera System for Bud Detection in Muscadine Grapevines. Agriculture, 15(21), 2220. https://doi.org/10.3390/agriculture15212220

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop