Next Article in Journal
Performance Optimization of Nonorthogonal MFSK for Symbol-by-Symbol Coherent Detection
Previous Article in Journal
Interface Quality Control of Self-Assembled Monolayer for Highly Sensitive Protein Detection Based on EGOFETs
Previous Article in Special Issue
Real-Time Non-Uniformity Correction Method for 800 FPS High-Frame-Rate Short-Wave Infrared Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Hierarchical YOLO-SAM: A Scalable Pipeline for Automated Segmentation and Morphometric Tracking of Coral Recruits in Time-Series Microscopy

1
Department of Computer Science, University of North Carolina at Chapel Hill, Chapel Hill, NC 27599, USA
2
Department of Mathematics and Statistics, University of North Carolina Wilmington, Wilmington, NC 28403, USA
3
Department of Biology and Marine Biology, Center for Marine Science, University of North Carolina Wilmington, Wilmington, NC 28403, USA
*
Authors to whom correspondence should be addressed.
Sensors 2026, 26(8), 2291; https://doi.org/10.3390/s26082291
Submission received: 24 February 2026 / Revised: 31 March 2026 / Accepted: 31 March 2026 / Published: 8 April 2026
(This article belongs to the Special Issue Digital Image Processing and Sensing Technologies—Second Edition)

Abstract

Coral reef ecosystems are declining rapidly due to climate change, disease, and anthropogenic stressors, driving the expansion of land-based coral propagation for reef restoration. A major bottleneck in these efforts is the manual measurement of coral recruit tissue area from microscopy images, which requires 2–7 min per image and limits scalability. We present a hierarchical deep learning pipeline that automates this measurement by integrating YOLO-based detection with Segment Anything Model (SAM) segmentation. YOLO localizes recruits and classifies them by developmental stage; stage-specific fine-tuned SAM models then segment live tissue using bounding box and background point prompts to suppress segmentation leakage and improve boundary precision. Surface area is computed directly from the segmented masks using pixel size extracted from image metadata. The pipeline reduces processing time to approximately 3–5 s per image—a 24–140× speedup over manual tracing. Evaluated on 3668 microscopy images from two national coral research facilities, the system achieves a mean IoU exceeding 95% and an auto-acceptance rate (AAR) of 71.51%, where predicted-to-ground-truth area ratios fall within a ±5% tolerance of expert annotation, substantially reducing manual workload while maintaining measurement reliability across species, developmental stages, and imaging conditions. This workflow addresses a critical bottleneck in restoration research and demonstrates the broader applicability of AI-based image analysis in marine ecology.

1. Introduction

Coral reefs are among the most biodiverse ecosystems on Earth, providing multiple ecosystem services and playing a key role in global economies [1,2,3]. However, global coral cover has declined by approximately 50% since 1957 [4,5]. Climate change is the main driver of this decline, with increased sea surface temperatures causing global coral bleaching events that are increasing in frequency and severity [6,7] Although local management efforts exist, they are largely ineffective against such global stressors, making restoration and intervention strategies an essential pillar of reef conservation [8].
Coral recruitment, the process in which coral larvae settle and metamorphose on substrates or existing reefs, is vital to restoration efforts [9]. Ex situ spawning facilities simulate temperature and light cycles to spawn corals for research and restoration efforts. Some of these facilities are still in their infancy, and researchers are working to understand how to optimize coral production by understanding the factors (e.g., substrate material, water flow, lighting, nutrition) that affect recruit health and growth [10,11,12,13]. These facilities measure recruit growth over time by manually tracing the polyp in microscopy images and using software to calculate the surface area [14,15]. However, manual tracing is slow (2 to 7 min per image) and labor intensive, limiting its scalability for large-scale restoration trials [16,17]. Traditional machine learning methods have been used to segment coral tissue, but struggle to generalize between datasets with varying lighting conditions, substrate types, species, and coral morphologies [18]. Other automated systems, often designed for reef surveys or remote sensing, lack the resolution and accuracy for millimeter-scale analysis of juvenile coral [18,19,20,21,22,23,24].
To overcome these challenges, we propose a hybrid framework that integrates YOLO (You Only Look Once) for efficient region-of-interest detection with SAM (Segment Anything Model). Note that YOLO and SAM are used throughout as general terms for their respective model families, unless otherwise specified. By unifying high-throughput detection with precise, annotation-free pixel-level segmentation, this approach eliminates the need for costly manual labeling. this approach significantly reduces the need for costly manual labeling. Unlike traditional machine learning methods, deep learning models automatically extract hierarchical features from data [25,26], enabling robust generalization across complex imaging conditions—a quality that has proven effective in medical imaging tasks requiring high precision in biological structures [27].
We further contribute three targeted improvements to enhance segmentation accuracy and biological relevance:
  • Stage-specific fine-tuning. Two dedicated SAM models are trained separately for early- and late-stage coral recruit development, directly addressing the distinct morphological and imaging challenges of each developmental phase.
  • Prompt engineering with negative prompts. Background points are incorporated as negative prompts alongside standard bounding box inputs, constraining mask expansion—a critical refinement for early-stage recruits where boundary ambiguity is greatest.
  • End-to-end evaluation. We present a comprehensive assessment of the full pipeline, demonstrating that automated, high-throughput analysis can match or exceed expert annotation accuracy while reducing processing time by one to two orders of magnitude (24–140×).
Together, these contributions establish a reproducible and scalable framework for coral recruit monitoring across large datasets, offering a practical path toward automating reef assessment at ecologically meaningful scales.

2. Materials

2.1. Data Collection and Background

This study used coral recruit images from two research facilities: the Coral REEF Laboratory at the University of North Carolina Wilmington (UNCW) and the Mote Marine Laboratory and Aquarium (Mote). Although there are slight differences in equipment and procedures employed by the two laboratories, both use high-resolution microscopes to capture juvenile coral recruits that are difficult to detect with the naked eye. At UNCW, images were taken using an Olympus LC30 camera (Olympus Corporation, Tokyo, Japan). mounted on an SZ61 stereo microscope (Olympus Corporation, Tokyo, Japan). This was used in tandem with the cellSens imaging software (cellSens Entry version, Olympus Corporation, Tokyo, Japan), allowing for embedded pixel-to-micrometer conversion metadata in the final TIFF image.
Coral recruit growth images from three projects were analyzed. Table 1 showed that the datasets span three cohorts collected at two laboratories over two years and include four coral species. UNCW datasets (2023–2024) comprise of Pseudodiploria clivosa (PCLI) and Montastraea cavernosa (MCAV) recruits photographed at seven time points at approximately 4-week intervals, with image counts declining from ∼400 to ∼200 as recruits died or were removed. The Mote dataset (2024) includes Colpophyllia natans (CNAT) and Pseudodiploria strigosa (PSTR) recruits photographed at four time points at approximately 8-week intervals, with image counts declining from ∼600 to ∼400.
Approximately one-third of the images were used for YOLO-SAM model development and split into training, validation, and test sets with a ratio of (70:20:10). Performance was evaluated systematically using the two most recent datasets (UNCW 2024 and Mote 2024) with a total of 3668 images. Together, these datasets capture variability in species composition, temporal sampling, and laboratory imaging workflows, enabling assessment of the deep learning pipeline’s robustness and cross-laboratory generalizability.
With this data, a qualitative classification system based on tissue coloration was provided. The images of coral recruits were assigned to categories A-G according to tissue coloration, growth maturity, and visual development (see Figure 1). This was necessary since coral growth rates varied between tanks, making comparisons by time point less effective. This qualitative classification allowed the pipeline to be evaluated based on visual distinction, rather than just by weeks. Furthermore, having a qualitative classification system made it easier to define methodologies to annotate images. Specifically, stage A–C is defined as early-stage, while stage D–G is defined as late-stage. Moreover, Figure 2 shows representative examples of each qualitative classification from different species/laboratories, lighting conditions, and backgrounds.

2.2. Image Preparation and Data Augmentation

The training dataset was restricted to microscopy images containing a single, isolated coral recruit, consistent with the study’s focus on individual-level detection, segmentation, and live tissue measurement. Images were excluded if they contained overlapping or adjacent recruits, physical damage, septal overgrowth, or imaging artifacts—specifically overexposure, underexposure, blur, or uneven illumination—identified through systematic quality screening prior to annotation. Although this single-recruit constraint reflects the scope of available ground-truth annotations rather than an architectural limitation, the framework is extensible to multi-recruit images, requiring only annotated training data with individual bounding boxes and segmentation masks per recruit.
Data augmentation was applied to the training images to improve generalization and prevent overfitting, following best practices in deep learning image analysis [28,29]. Training data was augmented threefold using rotations, grayscale conversion, and brightness and saturation adjustments, applied during both YOLO training and SAM fine-tuning to improve robustness without increasing model complexity or inference cost. Geometric augmentations included horizontal and vertical flips, random rotations (±15°), 90° rotations, and random cropping (<20% zoom). Color and noise augmentations comprised grayscale conversion (15% of images), hue shifts (±10°), saturation and brightness adjustments (±20%), exposure changes (±15%), Gaussian blurring (<2.5 pixels), and random noise injection (0.1% of pixels). Images were resized to meet the requirements of each model: 640 × 640 pixels for YOLO, where coarse features suffice for recruit detection, and 1024 × 1024 pixels for SAM, where higher resolution is needed for precise tissue segmentation.

2.3. Coral Recruit Annotation

Ground-truth surface areas were generated for the two recent growth projects (UNCW 2024 and Mote 2024) to complete the full AI system evaluation. Annotation guidelines provided by experts from the UNCW Coral REEF Laboratory informed a standardized polygon tracing scheme, implemented in cellSens, that delineated only the live tissue (polyp) of each coral recruit while excluding the corallite cup (Figure 3).

3. Methods

3.1. Hierarchical Developmental-Stage Architecture

In Figure 3, the coral microscopy images exhibit pronounced, developmentally driven heterogeneity in tissue morphology, symbiont pigmentation, and tissue–skeleton contrast across post-settlement stages. Early-stage recruits showed thin, weakly pigmented tissue confined to the skeletal cup and diffuse boundaries, and late-stage recruits displayed thicker, continuous tissue with clearer skeletal separation.
In our preliminary studies, initial fine-tuning of a single SAM model using a mixed dataset of early-stage and late-stage coral recruits resulted in degraded segmentation performance, particularly for early-stage corals, where segmentation masks expanded beyond the biologically relevant tissue region. This behavior was attributed to stage-dependent boundary semantics: early-stage coral recruits require segmentation confined to the upper tissue layer, whereas late-stage corals require outward expansion to capture thicker tissue and skeletal structure.
To address these challenges, we proposed a hierarchical architecture that explicitly accounts for the coral developmental stage. First, a YOLO-based detector was trained with an additional class label to classify coral recruits into developmental categories (e.g., early- vs. late-stage). Secondly, two stage-specific SAM models were fine-tuned independently using only images from their corresponding developmental stage, ensuring consistent boundary supervision within each model. During inference, each image is first classified by developmental stage, and then the corresponding stage-specific SAM model is applied for segmentation. This hierarchical approach reduces conflicting supervision during fine-tuning and produces segmentation masks that are more consistent with the biological morphology of coral recruits throughout developmental stages.
The proposed architecture (Figure 4) consists of an object detection model based on YOLO to detect and classify coral recruits, and two SAM-based segmentation models that delineate coral tissue for early or late coral images. The YOLO model performs detection to generate a bounding-box prompt that indicates regions of interest and classification to identify the coral recruit as early- or late-stage. This allows for the correct SAM to be chosen, while also guiding the SAM to focus on regions of interest for more detailed segmentation. If the coral recruit is classified as early-stage, further prompt engineering is used by adding negative prompts at each side of the bounding box to greater restrict the SAM.

3.2. YOLO

YOLO (You Only Look Once) is a single-stage object detector that performs localization and classification in a single forward pass. Its architecture comprises a convolutional backbone for hierarchical feature extraction, a neck for multi-scale feature aggregation, and a detection head that outputs bounding box coordinates and class probabilities. This design balances inference speed with detection accuracy across object sizes, making it well suited for high-volume image processing workflows. In this study, the YOLO module detects and classifies coral recruits in microscopy images, where multi-scale detection is particularly advantageous for small or low-contrast targets [30]. YOLO generates bounding boxes used as prompts for SAM and classifies recruits as early- or late-stage, enabling automatic selection of the appropriate SAM model. Unlike SAM, YOLO is pretrained on general object categories (COCO) and requires fine-tuning on domain-specific data for reliable coral recruit detection.
The YOLO module was iteratively refined to improve detection, particularly for early-stage recruits with poorly defined tissue boundaries. Initial YOLOv8 [31] trials produced occasional bounding box misalignments that propagated localization errors to downstream segmentation. Upgrading to YOLOv11 [32] improved robustness under high background clutter and low-contrast conditions. Grayscale preprocessing was evaluated to reduce color dependence but yielded no improvement over RGB, confirming that chromatic cues are biologically informative. For the final model, YOLOv11 was fine-tuned on a labeled dataset of coral recruit microscopy images from two research laboratories, augmented with images from the Mote growth project to improve diversity across growth stages and imaging conditions. Recruits were annotated with bounding boxes and classified as early- or late-stage. The dataset was split into 70% training, 20% validation, and 10% testing, using default Ultralytics hyperparameters except for 50 epochs and a batch size of 4 due to GPU memory constraints, with automatic mixed precision (AMP) enabled. Given the high detection accuracy achieved, recruit localization was treated as a resolved component and held fixed for subsequent pipeline development.

3.3. SAMs

SAM’s encoder–decoder design processes images and prompts (e.g., bounding boxes) independently before combining them to produce binary segmentation masks, enabling strong zero-shot generalization through pretraining on 11 million images [33]. When a specific version is referenced, SAM ViT-H denotes the original model [33] and SAM2-Hiera-Large denotes the second-generation model [34]; the two differ principally in backbone architecture, with SAM ViT-H using a plain Vision Transformer encoder and SAM2-Hiera-Large adopting a hierarchical Hiera encoder with a larger attention-head count and reduced checkpoint size.
In our pipeline, YOLO-derived bounding boxes and negative prompts guide SAM to segment live coral tissue from surrounding corallite, skeleton, and substrate. Although SAM generalizes well in zero-shot settings, initial performance on early-stage recruits was inconsistent due to indistinct tissue boundaries and low contrast. To address this, we fine-tuned two separate SAM models—one for each developmental stage—using manually annotated segmentation masks. Early-stage fine-tuning focused on reducing over-segmentation into the corallite cup, while late-stage fine-tuning improved boundary delineation along the tissue edge. Both models were trained using YOLO-derived bounding boxes as prompts and manual segmentation masks as ground truth, for 40 epochs with a batch size of 4, a learning rate of 1 × 10 5 , and a weight decay of 4 × 10 5 to preserve pretrained weights during adaptation. These stage-specific models replaced the default SAM checkpoint in the final pipeline. All model development and training were conducted on cloud-based GPU resources (Google Colab). Both stage-specific SAM2-Hiera-Large models were trained for up to 40 epochs with early stopping; validation loss plateaued at approximately 18 and 25 epochs for early-stage and late-stage SAM models, respectively. Fine-tuning was performed on NVIDIA A100 GPUs (∼7 min per epoch; ∼2–3 h per model). In contrast, training on NVIDIA T4 GPUs required ∼20 min per epoch. All inference experiments were conducted on NVIDIA T4 GPUs.
Segmentation performance was further improved through prompt engineering. Beyond bounding boxes, we added four background points at the midpoint of each bounding-box edge as negative prompts to suppress mask expansion into non-tissue regions—particularly the corallite cup and surrounding substrate in early-stage recruits. Each bounding box is defined by its corner coordinates ( x min , y min ) and ( x max , y max ) , recorded as absolute pixel coordinates alongside the source image dimensions; four background points are placed at the midpoints of each edge— ( x c , y min ) , ( x c , y max ) , ( x min , y c ) , and ( x max , y c ) —where x c = x min + x max 2 and y c = y min + y max 2 . As shown in Figure 5, adding four background (negative) points at the midpoints of the bounding-box edges (Figure 5a) reduced mask leakage and improved tissue boundary delineation under low-contrast conditions where tissue and cup are difficult to distinguish, compared to bounding-box prompts alone (Figure 5b). This was especially critical given that small boundary errors can propagate into large relative errors in downstream area estimates.
We evaluated three SAM variants—SAM ViT-H [33], SAM2-Hiera-Base-Plus, and SAM2-Hiera-Large [34]—selecting SAM2-Hiera-Large for its greater number of attention heads and superior negative prompt utilization. SAM2’s Hiera encoder achieves comparable representational capacity to ViT-H with fewer parameters, and prior evaluations confirm it matches or exceeds SAM ViT-H performance under proper calibration. Combined with fine-tuning and stage-specific model selection, these strategies adapted SAM2 into a high-precision segmentation module for coral morphology. Hereafter, SAM2 refers specifically to SAM2-Hiera-Large unless otherwise stated. Stage-specific SAM2 configurations are summarized in Table 2.

3.4. Performance Evaluation Metrics

Pipeline performance was evaluated at two levels. Standard machine learning metrics assessed YOLO and SAM individually during training. End-to-end evaluation then compared three configurations—prototype, optimized, and class-specific model switching in an ablation study—to quantify the effects of prompt design, fine-tuning, and stage-based model selection on auto-acceptance rate (AAR) and measurement accuracy.

3.4.1. Component-Level Evaluation

During fine-tuning, YOLO and SAM were evaluated on held-out validation sets to guide model selection. YOLO was assessed using precision, recall, and mean average precision at IoU threshold 0.5 (mAP@50), capturing localization accuracy and detection reliability. SAM was evaluated using Intersection over Union (IoU) between predicted and manually traced ground-truth masks, which penalizes both over- and under-segmentation and is sensitive to boundary misalignment—critical for area estimation in growth analysis. Loss and performance curves were tracked across epochs to monitor learning stability. Additionally, both models were evaluated on a held-out test set to confirm generalization and rule out overfitting or data leakage.

3.4.2. Pipeline-Level Evaluation

Three pipeline configurations were evaluated to quantify developmental improvements. The prototype used YOLOv8, SAM ViT-H, box-only prompts, and no fine-tuning, serving as a baseline control. The final configuration used YOLOv11, SAM2-Hiera-Large, bounding box and background prompts, and stage-specific fine-tuned SAM models, representing the final optimized pipeline. A third class-specific configuration inverted the model–stage assignments to validate that each fine-tuned SAM performed best on its intended recruit stage, confirming the dual-branch design. Configuration differences are summarized in Table 3. All configurations were evaluated on the UNCW 2024 and Mote 2024 coral growth datasets. Unlike component-level evaluation, which targeted segmentation and detection accuracy (IoU, mAP@50), pipeline-level evaluation assessed measurement accuracy—how closely predicted recruit areas matched ground-truth measurements—using multiple statistical indicators at each time point.
Surface Area Quantification
The real-world surface area ( A real ) was computed directly from the segmented masks using pixel size extracted from the image metadata, as follows:
A real = A px × s 2 ,
where A px denotes the number of pixels within the segmented mask and s the pixel size (real-world length per pixel). When the pixel size is provided in micrometers per pixel (μm/pixel), the surface area is obtained in square micrometers (μm2).
Area Ratio (AR)
The area ratio quantifies systematic bias in a predicted area relative to ground truth:
A R i = A i pred A i true , i = 1 , , N ,
where A i pred and A i true are the predicted and manually traced areas for the i-th image, and N is the total number of images at a given time point. AR = 1 indicates perfect agreement; AR > 1 and AR < 1 indicate over- and underestimation, respectively.
Auto-Acceptance Rate (AAR)
The AAR is the proportion of predictions within 5% of the ground truth area:
AAR = 1 N i = 1 N 1 A R i 1 < 0.05 × 100 ,
where 1 ( · ) is an indicator function equal to one when A R i [ 0.95 , 1.05 ] and zero otherwise. Metrics were computed at each time point across both datasets to capture temporal variability.

4. Results

To validate individual components and fine-tuning effects, we evaluated detection and segmentation modules prior to full pipeline testing, ensuring each module achieved the precision required for accurate coral recruit measurement.

4.1. Object Detection Performance (YOLO)

Table 4 reports YOLOv11 detection performance on the held-out test set. High precision, recall, and mAP@50 across both stages indicate robust recruit localization suitable for downstream segmentation. The lower mAP@50-95 for early-stage recruits (0.873 vs. 0.948) reflects their diffuse tissue margins and reduced contrast against the skeleton, which makes precise bounding-box alignment more challenging at stricter IoU thresholds.
Training curves (Figure 6) show stable convergence over 50 epochs without overfitting, with validation metrics closely tracking training performance—indicating effective generalization across species, lighting conditions, and substrate types in the dataset.

4.2. Segmentation Performance (SAM)

Both stage-specific SAM models achieved high segmentation accuracy (Table 5). The late-stage model reached an IoU of 0.9760 and the early-stage model 0.9598. The early-stage model achieved a lower final loss (0.0025 vs. 0.0100) but slightly lower IoU, consistent with the greater ambiguity at early-stage tissue boundaries where live tissue and corallite cup are difficult to distinguish. The late-stage model’s higher loss but superior IoU reflects successful adaptation to more complex but better-defined tissue morphologies.
Training curves (Figure 7) show rapid convergence and stable validation loss for both models. The distinct convergence patterns reflect stage-specific segmentation characteristics and support the dual-model strategy over a single general-purpose model.

4.3. Pipeline Performance Across Developmental Stages

The baseline used YOLOv8 with a pretrained SAM ViT-H and bounding-box prompts only. The final pipeline incorporated YOLOv11, dual stage-specific fine-tuned SAM2-Hiera-Large models, and bounding box with four background point prompts. Both were evaluated on the UNCW 2024 and Mote 2024 datasets (Table 6, Figure 8).
The baseline performed reliably on late-stage recruits, achieving an AAR of 65–92% at later time points with median area ratios near unity (0.96–0.99), indicating slight but consistent underestimation. Early-stage performance was substantially weaker, with an AAR of 6–59% driven by overestimation from tissue transparency and illumination artifacts. This was most pronounced in the Mote dataset, where segmentation leakage into the corallite cup produced median area ratios of 1.04–1.21. Intermediate time points in the UNCW dataset (weeks 8–12), containing mixed stages, yielded a moderate AAR of 34–42%. Mote week 8, which visually resembled late-stage recruits, achieved a higher AAR (75%) but retained early-stage-like overestimation (median area ratio ≈ 1.03).
The final pipeline improved performance across most time points. Gains were modest for late-stage recruits (AAR +3–10%) but substantial for early-stage recruits, where AAR increased by 20–54% and median area ratios shifted from 1.04–1.21 to 0.96–1.00, reflecting effective suppression of overestimation. At intermediate time points, the AAR improved by 12–13% in UNCW but declined moderately at Mote week 8 (−7.35%), where the median area ratio increased marginally from 1.03 to 1.04.

4.4. Ablation: Stage-Inverted Processing

To verify the dual-branch design, we conducted a stage-inversion experiment: early-stage images were processed with the late-stage model and configuration (no background prompts), and late-stage images with the early-stage model and configuration (with background prompts).
As shown in Table 7 and Figure 9, inversion degraded performance at 8 of 10 time points, with AAR reductions of up to 46.63% (UNCW week 18). The two exceptions both occurred at developmentally ambiguous time points: UNCW week 4 (early, ΔAAR = + 3.79 %) and Mote week 8 (early/late, ΔAAR = +21.38%). The Mote week 8 anomaly is notable—the inverted late-stage configuration outperformed the correct assignment, with median area ratios shifting from 1.036 to 1.003, suggesting these recruits were functionally closer to late-stage morphology despite their mixed-stage classification. These two exceptions aside, the consistent performance degradation under inversion confirms that the stage-specific model assignment is mechanistically justified. The dual-branch design performs correctly as designed for the majority of the developmental range (8 of 10 time points, with degradations up to 46.6%), but performance at morphologically ambiguous transition points motivates future work on finer-grained stage resolution.
Additional visualization and expanded descriptive statistics can be found in Supplementary Materials. Detailed descriptive statistics for the Area Ratio across all pipelines and time points are provided in Table S1, and additional visualizations are presented in Figures S1–S4 of the Supplementary Materials. Figure S1 summarizes AVG Area Ratio (±STD) for the Baseline, Final, and Ablation pipelines across developmental stages at both sites. Figure S2 provides a heatmap overview of AVG AR, Median AR, and AAR across pipelines and time points. Figures S3 and S4 display the full AR scatter plots and boxplots for the UNCW and Mote sites, respectively, confirming that the Final pipeline consistently compresses AR distributions toward unity with reduced bias and variability across most time points.

5. Discussion

This study presents a deep learning pipeline for automated coral recruit area measurement that integrates YOLO-based detection with SAM-based segmentation. Processing time was reduced from 2 to 7 min per image by manual tracing to approximately 3 to 5 s per image on T4 GPUs and around 3 s on A100 GPUs under typical conditions, a 24–140× speedup that removes a major analytical bottleneck and enables large-scale restoration studies previously constrained by labor.
The optimized pipeline achieved a mean IoU exceeding 95% and an overall AAR of 71.51%, where predicted-to-ground-truth area ratios fall within a ±5% tolerance of expert annotation. The remaining 28.49% fell outside the ±5% tolerance threshold; however, as shown in the Supplementary Materials, most predictions clustered closely around unity, indicating marginal deviations rather than systematic failures. Predicted segmentations remained biologically reasonable across both datasets with minimal correction required. This shifts the analytical burden from full manual tracing to lightweight quality control, preserving measurement reliability at substantially reduced effort.
A key strength is generalization across datasets collected under distinct experimental conditions. Despite differences in imaging protocols, substrate types, water treatments, coral species, and developmental stages, performance remained consistent across the UNCW and Mote datasets. This suggests the pipeline captures fundamental visual features of coral tissue rather than overfitting to site-specific artifacts, supporting deployment across multiple laboratories and restoration programs.
The hierarchical stage-specific architecture directly addresses the biological challenge of rapid morphological change during early development. Early-stage recruits, characterized by translucent tissue and poorly defined tissue–skeleton boundaries, caused over-segmentation in generic models; incorporating a specialized early-stage SAM with background point prompts effectively suppressed segmentation leakage. For late-stage recruits, where tissue boundaries are more distinct, the late-stage model achieved consistently strong performance with only slight underestimation. Together, these design choices extend reliable measurement across the full recruitment trajectory. As shown in Figure 10, representative challenging cases illustrate the practical performance limits imposed by early developmental morphology, symbiont contrast, and illumination effects.
Several limitations warrant consideration. Pipeline accuracy depends on reliable YOLO detections, and extreme imaging conditions or irregular growth patterns can degrade bounding-box localization and propagate errors downstream. SAM fine-tuning requires substantial computational resources, which may limit accessibility for groups without modern hardware. Future work could explore lightweight segmentation architectures or transfer learning strategies that reduce computational demands while preserving accuracy. The A–G developmental stage classification was derived through systematic expert review of hundreds of recruit images and two-researcher consensus, and retains an inherent element of subjectivity. Future work should explore automating stage assignment using quantitative criteria (e.g., size, shape, or symbiont acquisition metrics) and establishing formal inter-rater reliability (e.g., Cohen’s kappa) to improve reproducibility and reduce manual intervention.
Although the pipeline performed well across four reef-building coral species over 0–23 weeks, validation on a broader taxonomic range and longer temporal scales remains necessary to confirm generalizability. Future studies could quantify tissue–substrate boundary contrast and intensity gradients across developmental stages, which would allow direct measurement of the biological contribution to segmentation uncertainty and help disentangle biological from algorithmic performance limits. A fully factorial ablation isolating the independent contributions of YOLO version, SAM architecture, and stage-specific configuration was outside the scope of this deployment-driven study; systematic component-wise evaluation represents a valuable direction for future work.
Object detection architectures have evolved rapidly since YOLOv8 [31], with subsequent iterations introducing improvements in accuracy, inference speed, and robustness under challenging imaging conditions. YOLOv11 [32] improved detection under clutter and low-contrast boundaries, directly motivating its adoption in our pipeline for early-stage coral recruit detection. More recent architectures—including transformer-based detectors and the next generation of foundation segmentation models—further advance detection precision and boundary delineation; however, as these models were released during the final stages of this work and the field continues to evolve rapidly, systematic benchmarking within coral morphometric pipelines remains an open research direction that we prioritize for future work.
The modular pipeline design extends naturally beyond coral recruits—with modest modifications, similar workflows could support larval settlement monitoring in oysters, sponges, or algae. Future directions include real-time deployment in coral nurseries, volumetric measurement via 3D imaging, cloud-based interfaces for broader community access, and active or self-supervised learning to reduce annotation burden in long-term, multi-environment monitoring. By automating recruit measurement with high accuracy and efficiency, this pipeline enables more scalable and standardized restoration research. As reef restoration efforts expand globally, tools that reduce manual labor while maintaining biological rigor will be critical for timely conservation decisions. Consistent with this potential, the pipeline has been operational at multiple coral research facilities since March 2025, supporting routine monitoring beyond the validation setting and demonstrating robustness to cross-institutional variation in imaging conditions. All participating laboratories run the pipeline via Google Colab with GPU runtime, which provides accessible T4 or A100 GPUs without requiring local hardware infrastructure.

6. Conclusions

We present a deep learning pipeline that automates coral recruit detection and surface area measurement in microscopy images, reducing processing time by 24–140× relative to manual tracing while achieving a mean IoU exceeding 95% and an overall AAR of 71.51%. Stage-specific SAM models and background point prompts were critical for handling morphological differences between early- and late-stage recruits, and the system generalized consistently across species, datasets, and imaging conditions. Remaining limitations include computational demands for fine-tuning and the need for validation across a broader taxonomic range.
As reef restoration efforts scale globally, automated tools that reduce manual labor while maintaining biological rigor are increasingly necessary. The modular design of this pipeline makes it adaptable to related ecological imaging tasks, and future extensions—including real-time deployment and 3D volumetric measurement—could further expand its utility for large-scale coral monitoring.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/s26082291/s1, Table S1: Descriptive statistics of Baseline, Final, and Ablation Pipeline area ratio across time points for UNCW and Mote sites; Figure S1: AVG Area Ratio (± STD) across developmental stages for UNCW (Weeks 0, 4, 8, 12, 18, 23) and Mote (Weeks 0, 8, 15, 21) sites: (a) Baseline vs. Final pipeline; (b) Final vs. Ablation pipeline; Figure S2: Heatmap summary of pipeline performance across time points for UNCW (Weeks 0, 4, 8, 12, 18, 23) and Mote (Weeks 0, 8, 15, 21): (a) AVG Area Ratio; (b) Median Area Ratio; (c) Auto-Accept Rate (%); Figure S3: Area Ratio distributions across all time points for the UNCW site (Weeks 0, 4, 8, 12, 18, 23) for Baseline, Final, and Ablation pipelines; Figure S4: Area Ratio distributions across all time points for the Mote site (Weeks 0, 8, 15, 21) for Baseline, Final, and Ablation pipelines.

Author Contributions

Conceptualization, R.S.Z., C.C. and M.V.H.; methodology, R.S.Z., C.C., M.V.H. and N.D.F.; software, R.S.Z. and C.C.; validation, R.S.Z. and C.C.; formal analysis, R.S.Z. and C.C.; investigation, R.S.Z., C.C., M.V.H. and N.D.F.; resources, M.V.H. and N.D.F.; data curation, M.V.H. and R.S.Z.; writing—original draft preparation, R.S.Z., C.C. and M.V.H.; writing—review and editing, C.C., R.S.Z., M.V.H. and N.D.F.; visualization, R.S.Z., C.C. and M.V.H.; supervision, C.C. and N.D.F.; project administration, C.C., M.V.H. and N.D.F.; funding acquisition, N.D.F. and C.C. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the NOAA’s Ruth Gates Coral Restoration Innovation Grant (NA22NMF4820207) and the Florida Department of Environmental Protection’s Coral Protection and Restoration Program (CPRP_C5B5BD).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Selected components of this project may be made available by the corresponding authors upon reasonable request, subject to institutional agreements and data-use policies.

Acknowledgments

The authors thank Simon Zhao and Eric Su for their assistance with image annotation and testing during this project. We thank Tyler Pratt for tracing many coral recruit images used to train the model. We thank Ashley Weeks for helping take recruit microscopy images. We thank Celia Leto at Mote Marine Laboratory for providing recruit images. During the preparation of this manuscript/study, the authors used ChatGPT (OpenAI, GPT-5.2), Claude (Anthropic, Claude 4), and DeepSeek (DeepSeek-V3) for generating text and language editing. The authors have reviewed and edited the generated content and take full responsibility for the content of this publication.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
YOLOYou Only Look Once
SAMSegment Anything Model
AARAuto-acceptance rate
IoUIntersection over Union
PCLIPseudodiploria clivosa
MCAVMontastraea cavernosa
CNATColpophyllia natans
PSTRPseudodiploria strigosa
MoteMote Marine Laboratory and Aquarium

References

  1. Eddy, T.D.; Lam, V.W.; Reygondeau, G.; Cisneros-Montemayor, A.M.; Greer, K.; Palomares, M.L.D.; Bruno, J.F.; Ota, Y.; Cheung, W.W. Global decline in capacity of coral reefs to provide ecosystem services. One Earth 2021, 4, 1278–1285. [Google Scholar] [CrossRef]
  2. Moberg, F.; Folke, C. Ecological goods and services of coral reef ecosystems. Ecol. Econ. 1999, 29, 215–233. [Google Scholar] [CrossRef]
  3. Spalding, M.; Burke, L.; Wood, S.A.; Ashpole, J.; Hutchison, J.; zu Ermgassen, P. Mapping the global value and distribution of coral reef tourism. Mar. Policy 2017, 82, 104–113. [Google Scholar] [CrossRef]
  4. Bruno, J.F.; Valdivia, A. Coral reef degradation is not correlated with local human population density. Sci. Rep. 2016, 6, 29778. [Google Scholar] [CrossRef] [PubMed]
  5. Field, C.B.; Barros, V.R.; Dokken, D.J.; Mach, K.J.; Mastrandrea, M.D.; Bilir, T.E.; Chatterjee, M.; Ebi, K.L.; Estrada, Y.O.; Genova, R.C.; et al. (Eds.) Climate Change 2014: Impacts, Adaptation, and Vulnerability. Part A: Global and Sectoral Aspects; Cambridge University Press: Cambridge, UK; New York, NY, USA, 2014. [Google Scholar]
  6. Hughes, T.P.; Kerry, J.T.; Álvarez-Noriega, M.; Álvarez-Romero, J.G.; Anderson, K.D.; Baird, A.H.; Babcock, R.C.; Beger, M.; Bellwood, D.R.; Berkelmans, R.; et al. Global warming and recurrent mass bleaching of corals. Nature 2017, 543, 373–377. [Google Scholar] [CrossRef]
  7. Hoegh-Guldberg, O.; Anthony, K.; Berkelmans, R.; Dove, S.; Fabricius, K.; Lough, J.; Marshall, P.; van Oppen, M.; Negri, A.; Willis, B. Vulnerability of Reef-Building Corals on the Great Barrier Reef to Climate Change; The University of Queensland: St Lucia, Australia, 2007. [Google Scholar]
  8. Voolstra, C.R.; Peixoto, R.S.; Ferrier-Pagès, C. Mitigating the ecological collapse of coral reef ecosystems. EMBO Rep. 2023, 24, e56826. [Google Scholar] [CrossRef]
  9. Edmunds, P.J. Coral recruitment: Patterns and processes determining the dynamics of coral populations. Biol. Rev. Camb. Philos. Soc. 2023, 98, 1862–1886. [Google Scholar] [CrossRef]
  10. Levenstein, M.A.; Marhaver, K.L.; Quinlan, Z.A.; Tholen, H.M.; Tichy, L.; Yus, J.; Lightcap, I.; Wegley Kelly, L.; Juarez, G.; Vermeij, M.J.A.; et al. Composite Substrates Reveal Inorganic Material Cues for Coral Larval Settlement. ACS Sustain. Chem. Eng. 2022, 10, 3960–3971. [Google Scholar] [CrossRef]
  11. Craggs, J.; Guest, J.R.; Davis, M.; Simmons, J.; Dashti, E.; Sweet, M. Inducing broadcast coral spawning ex situ: Closed system mesocosm design and husbandry protocol. Ecol. Evol. 2017, 7, 11066–11078. [Google Scholar] [CrossRef]
  12. Chamberland, V.F.; Vermeij, M.J.; Brittsan, M.; Carl, M.; Schick, M.; Snowden, S.; Schrier, A.; Petersen, D. Restoration of critically endangered elkhorn coral (Acropora palmata) populations using larvae reared from wild-caught gametes. Glob. Ecol. Conserv. 2015, 4, 526–537. [Google Scholar] [CrossRef]
  13. Khodzori, F.A.; Roger, N.A.A.; Nor’ashikin, A.Z.; Azseri, A.; Misi, L.L.; Mazni, M.A.; Hisham, H.K.; Shah, M.D.; Chong, W.S. Coral Aquaculture: A Review of In Situ and Ex Situ Culture Systems, Conditions, Applications, and Challenges. In Essentials of Aquaculture Practices; Faudzi, N.M., Shah, M.D., Mazlan, N., Raehanah Muhamad Shaleh, S., Eds.; Springer Nature: Singapore, 2024; pp. 239–265. [Google Scholar] [CrossRef]
  14. Zweifler, A.; Akkaynak, D.; Mass, T.; Treibitz, T. In situ Analysis of Coral Recruits Using Fluorescence Imaging. Front. Mar. Sci. 2017, 4, 273. [Google Scholar] [CrossRef]
  15. Bahr, K.D.; Tran, T.; Jury, C.P.; Toonen, R.J. Abundance, size, and survival of recruits of the reef coral Pocillopora acuta under ocean warming and acidification. PLoS ONE 2020, 15, e0228168. [Google Scholar] [CrossRef] [PubMed]
  16. Koch, H.R.; Wallace, B.; DeMerlis, A.; Clark, A.S.; Nowicki, R.J. 3D Scanning as a Tool to Measure Growth Rates of Live Coral Microfragments Used for Coral Reef Restoration. ProQuest 2021, 8, 623645. [Google Scholar] [CrossRef]
  17. O’Neil, K.L.; Serafin, R.M.; Patterson, J.T.; Craggs, J.R.K. Repeated ex situ spawning in two highly disease susceptible corals in the family Meandrinidae. Front. Mar. Sci. 2021, 8, 669976. [Google Scholar] [CrossRef]
  18. Macadam, A.; Nowell, C.J.; Quigley, K. Machine Learning for the Fast and Accurate Assessment of Fitness in Coral Early Life History. Remote Sens. 2021, 13, 3173. [Google Scholar] [CrossRef]
  19. Wong, Y.K.; Zheng, Z.; Zhang, M.; Suggett, D.; Yeung, S.K. CoralSCOP-LAT: Labeling and Analyzing Tool for Coral Reef Images with Dense Mask. arXiv 2024, arXiv:2410.20436. [Google Scholar] [CrossRef]
  20. Beijbom, O.; Edmunds, P.J.; Roelfsema, C.; Smith, J.; Kline, D.I.; Neal, B.P.; Dunlap, M.J.; Moriarty, V.; Fan, T.Y.; Tan, C.J.; et al. Towards Automated Annotation of Benthic Survey Images: Variability of Human Experts and Operational Modes of Automation. PLoS ONE 2015, 10, e0130312. [Google Scholar] [CrossRef]
  21. Richards, B.L.; Beijbom, O.; Campbell, M.D.; Clarke, M.E.; Cutter, G.; Dawkins, M.; Edington, D.; Hart, D.R.; Hill, M.C.; Hoogs, A.; et al. Automated Analysis of Underwater Imagery: Accomplishments, Products, and Vision; Technical Memorandum; U.S. Department of Commerce: Washington, DC, USA, 2019. [Google Scholar]
  22. González-Rivero, M.; Beijbom, O.; Rodriguez-Ramirez, A.; Bryant, D.E.P.; Ganase, A.; Gonzalez-Marrero, Y.; Herrera-Reveles, A.; Kennedy, E.V.; Kim, C.J.S.; Lopez-Marcano, S.; et al. Monitoring of Coral Reefs Using Artificial Intelligence: A Feasible and Cost-Effective Approach. Remote Sens. 2020, 12, 489. [Google Scholar] [CrossRef]
  23. Xu, M.; Su, J.; Liu, Y. AquaSAM: Underwater Image Foreground Segmentation. arXiv 2023, arXiv:2308.04218. [Google Scholar] [CrossRef]
  24. Chowdhury, A.; Jahan, M.; Kaisar, S.; Khoda, M.E.; Rajin, S.; Naha, R. Coral Reef Surveillance with Machine Learning: A Review of Datasets, Techniques, and Challenges. Electronics 2024, 13, 5027. [Google Scholar] [CrossRef]
  25. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  26. Guo, Y.; Liu, Y.; Oerlemans, A.; Lao, S.; Wu, S.; Lew, M.S. Deep learning for visual understanding: A review. Neurocomputing 2016, 187, 27–48. [Google Scholar] [CrossRef]
  27. Pandey, S.; Chen, K.F.; Dam, E.B. Comprehensive Multimodal Segmentation in Medical Imaging: Combining YOLOv8 with SAM and HQ-SAM Models. arXiv 2023, arXiv:2310.12995. [Google Scholar] [CrossRef]
  28. Perez, L.; Wang, J. The Effectiveness of Data Augmentation in Image Classification using Deep Learning. arXiv 2017, arXiv:1712.04621. [Google Scholar] [CrossRef]
  29. Shorten, C.; Khoshgoftaar, T.M. A survey on Image Data Augmentation for Deep Learning. J. Big Data 2019, 6, 60. [Google Scholar] [CrossRef]
  30. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You Only Look Once: Unified, Real-Time Object Detection. arXiv 2016, arXiv:1506.02640. [Google Scholar] [CrossRef]
  31. Jocher, G.; Chaurasia, A.; Qiu, J. Ultralytics YOLOv8. 2023. Available online: https://github.com/ultralytics/ultralytics (accessed on 30 March 2026).
  32. Jocher, G.; Qiu, J. Ultralytics YOLO11. 2024. Available online: https://docs.ultralytics.com/models/yolo11/ (accessed on 30 March 2026).
  33. Kirillov, A.; Mintun, E.; Ravi, N.; Mao, H.; Rolland, C.; Gustafson, L.; Xiao, T.; Whitehead, S.; Berg, A.C.; Lo, W.Y.; et al. Segment Anything. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV); IEEE: New York, NY, USA, 2023; pp. 4015–4026. [Google Scholar] [CrossRef]
  34. Ravi, N.; Gabeur, V.; Hu, Y.T.; Hu, R.; Ryali, C.; Ma, T.; Khedr, H.; Rädle, R.; Rolland, C.; Gustafson, L.; et al. SAM 2: Segment Anything in Images and Videos. arXiv 2024, arXiv:2408.00714. [Google Scholar]
Figure 1. Qualitative estimates of zooxanthellae density from Stage (AG), across Weeks 0 to 28 post-fertilization.
Figure 1. Qualitative estimates of zooxanthellae density from Stage (AG), across Weeks 0 to 28 post-fertilization.
Sensors 26 02291 g001
Figure 2. Example images from each stage (AG), highlighting tissue coloration, maturity, and growth progression.
Figure 2. Example images from each stage (AG), highlighting tissue coloration, maturity, and growth progression.
Sensors 26 02291 g002
Figure 3. Representative coral recruits across post-settlement developmental stages. Top row shows raw microscopy images and bottom row shows corresponding manual tissue annotations in cellSens (red). Stages progress from early recruits without symbionts (1), to symbiont-colonized and increasingly developed recruits with varying degrees of skeletal cup exposure (26), culminating in large recruits with complete tissue coverage of the skeletal cup (7).
Figure 3. Representative coral recruits across post-settlement developmental stages. Top row shows raw microscopy images and bottom row shows corresponding manual tissue annotations in cellSens (red). Stages progress from early recruits without symbionts (1), to symbiont-colonized and increasingly developed recruits with varying degrees of skeletal cup exposure (26), culminating in large recruits with complete tissue coverage of the skeletal cup (7).
Sensors 26 02291 g003
Figure 4. Hierarchical architecture of the proposed YOLO–SAM framework with stage-specific prompt strategies: YOLO performs object detection and stage classification. Early SAM uses bounding-box prompts with four additional background points as negative prompts, while Late SAM uses bounding-box prompts only.
Figure 4. Hierarchical architecture of the proposed YOLO–SAM framework with stage-specific prompt strategies: YOLO performs object detection and stage classification. Early SAM uses bounding-box prompts with four additional background points as negative prompts, while Late SAM uses bounding-box prompts only.
Sensors 26 02291 g004
Figure 5. Effect of background-point prompts on early-stage SAM2 segmentation: (a) with four background (negative) points placed at the midpoints of the bounding-box edges and (b) without background-point prompts. Green: YOLO bounding box; red: SAM2 segmentation contour.
Figure 5. Effect of background-point prompts on early-stage SAM2 segmentation: (a) with four background (negative) points placed at the midpoints of the bounding-box edges and (b) without background-point prompts. Green: YOLO bounding box; red: SAM2 segmentation contour.
Sensors 26 02291 g005
Figure 6. YOLOv11 training and validation performance across epochs.
Figure 6. YOLOv11 training and validation performance across epochs.
Sensors 26 02291 g006
Figure 7. Validation loss and IoU across epochs for early- and late-stage SAM models. Red: validation loss (left axis); blue: IoU (right axis).
Figure 7. Validation loss and IoU across epochs for early- and late-stage SAM models. Red: validation loss (left axis); blue: IoU (right axis).
Sensors 26 02291 g007
Figure 8. Auto-acceptance rate (AAR) improvement across developmental stages.
Figure 8. Auto-acceptance rate (AAR) improvement across developmental stages.
Sensors 26 02291 g008
Figure 9. Visual comparison between the inverted and final auto-acceptance rate (AAR).
Figure 9. Visual comparison between the inverted and final auto-acceptance rate (AAR).
Sensors 26 02291 g009
Figure 10. Representative challenging coral recruit microscopy images across developmental stages, including tissue transparency, weak contrast, skeletal cup interference, uneven illumination, symbiont-dominated morphology, and increasing structural complexity. Red contours show model-predicted segmentations and green boxes indicate detection bounding boxes. Visually plausible masks can still produce large area-ratio deviations due to biological ambiguity and imaging artifacts.
Figure 10. Representative challenging coral recruit microscopy images across developmental stages, including tissue transparency, weak contrast, skeletal cup interference, uneven illumination, symbiont-dominated morphology, and increasing structural complexity. Red contours show model-predicted segmentations and green boxes indicate detection bounding boxes. Visually plausible masks can still produce large area-ratio deviations due to biological ambiguity and imaging artifacts.
Sensors 26 02291 g010
Table 1. Overview of coral growth projects used in the study.
Table 1. Overview of coral growth projects used in the study.
LabYearSpeciesTime Points SampledApprox Images Declining
UNCW2023PCLI7 (every ∼4 weeks)400 → 200
UNCW2024MCAV7 (every ∼4 weeks)400 → 200
Mote2024CNAT, PSTR4 (every ∼8 weeks)600 → 400
Table 2. Stage-specific SAM2 segmentation configurations.
Table 2. Stage-specific SAM2 segmentation configurations.
ComponentEarly-Stage SAM2Late-Stage SAM2
SAM architectureSAM2-Hiera-Largesame
YOLO inputBounding box + stage labelsame
Bounding-box promptYessame
Background pointsFour (box-edge midpoints)None
Foreground pointsNonesame
Fine-tuned weightsEarly-stage specificLate-stage specific
Primary objectiveSuppress cup over-segmentationPreserve tissue boundaries
Typical tissue contrastLow/indistinctHigh/well-defined
Table 3. Comparison between the baseline (prototype) and final (optimized) YOLO–SAM pipeline configurations.
Table 3. Comparison between the baseline (prototype) and final (optimized) YOLO–SAM pipeline configurations.
ComponentBaseline (Prototype)Final (Optimized)
YOLO versionYOLOv8YOLOv11
YOLO outputBounding box onlyBounding box + stage classification
SAM architectureSAM ViT-H (pretrained)SAM2-Hiera-Large
SAM fine-tuningNoneDual stage-specific fine-tuned models
Stage-specific switchingNoYes (early vs. late routing)
Bounding-box promptYesYes
Background pointsNoneFour (box-edge midpoints, early only)
Segmentation strategySingle SAM modelStage-specific SAM2 models
Table 4. YOLOv11 detection performance on the test set (All = combined early + late; early/late = class-specific).
Table 4. YOLOv11 detection performance on the test set (All = combined early + late; early/late = class-specific).
ClassImages/
Instances
PrecisionRecallmAP@50mAP@50-95
All4170.9720.9860.9930.911
Early2080.9710.9860.9920.873
Late2090.9730.9860.9940.948
Table 5. Loss and IoU metrics for Early and Late SAMs on the test set.
Table 5. Loss and IoU metrics for Early and Late SAMs on the test set.
ModelLossIoU
Early SAM0.00250.9598
Late SAM0.01000.9760
Table 6. Performance comparison between baseline and final pipelines across datasets and developmental stages.
Table 6. Performance comparison between baseline and final pipelines across datasets and developmental stages.
DatasetWeekStageAARBaseAARFinalΔAARMARBaseMARFinalΔ(|Bias|)
UNCW0Early59.0383.33+24.301.0370.989+0.026
UNCW4Early28.9149.17+20.261.0380.964+0.002
UNCW8Early/Late33.9647.19+13.230.9590.954−0.005
UNCW12Early/Late41.4953.56+12.070.9650.955−0.010
UNCW18Late65.8775.96+10.090.9670.970+0.003
UNCW23Late84.2988.48+4.190.9790.980+0.001
Mote0Early5.8660.13+54.271.2050.995+0.200
Mote8Early/Late75.2867.93−7.351.0261.036−0.010
Mote15Late87.3693.10+5.740.9910.999+0.008
Mote21Late92.0095.29+3.290.9880.997+0.009
AAR: auto-acceptance rate, percentage of predictions within 5% of ground-truth area. ΔAAR = AARFinal − AARBase; positive values indicate improved auto-acceptance in the final pipeline. MAR: median of A R i = A i pred / A i true ; values near 1 indicate low systematic bias. Δ(|Bias|): reduction in absolute deviation from unity, | MAR Base 1 |     | MAR Final 1 | ; positive values indicate improved agreement with ground truth.
Table 7. Pipeline performance under inverted stage-model assignment, validating the dual-branch SAM design.
Table 7. Pipeline performance under inverted stage-model assignment, validating the dual-branch SAM design.
DatasetWeekStageAARFinalAARAblationΔAARMARFinalMARAblationΔ(|Bias|)
UNCW0Early83.3382.64−0.690.9891.013−0.002
UNCW4Early49.1752.96+3.790.9640.995+0.031
UNCW8Early/Late47.1927.72−19.470.9540.935−0.019
UNCW12Early/Late53.5626.78−26.780.9550.938−0.017
UNCW18Late75.9629.33−46.630.9700.945−0.025
UNCW23Late88.4870.68−17.800.9800.961−0.019
Mote0Early60.1341.37−18.760.9951.057−0.052
Mote8Early/Late67.9389.31+21.381.0361.003+0.033
Mote15Late93.1073.10−20.000.9990.966−0.033
Mote21Late95.2984.94−10.350.9970.971−0.026
AAR: auto-acceptance rate, percentage of predictions within 5% of ground-truth area. ΔAAR = AARAblation − AARFinal; negative values confirm that correct stage assignment outperforms inverted assignment. MAR: median of A R i = A i pred / A i true ; values near 1 indicate low systematic bias. Δ(|Bias|) = | MAR Final 1 | | MAR Ablation 1 | ; negative values indicate lower systematic bias in the final pipeline.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhao, R.S.; Chen, C.; Van Horn, M.; Fogarty, N.D. Hierarchical YOLO-SAM: A Scalable Pipeline for Automated Segmentation and Morphometric Tracking of Coral Recruits in Time-Series Microscopy. Sensors 2026, 26, 2291. https://doi.org/10.3390/s26082291

AMA Style

Zhao RS, Chen C, Van Horn M, Fogarty ND. Hierarchical YOLO-SAM: A Scalable Pipeline for Automated Segmentation and Morphometric Tracking of Coral Recruits in Time-Series Microscopy. Sensors. 2026; 26(8):2291. https://doi.org/10.3390/s26082291

Chicago/Turabian Style

Zhao, Richard S., Cuixian Chen, Meg Van Horn, and Nicole D. Fogarty. 2026. "Hierarchical YOLO-SAM: A Scalable Pipeline for Automated Segmentation and Morphometric Tracking of Coral Recruits in Time-Series Microscopy" Sensors 26, no. 8: 2291. https://doi.org/10.3390/s26082291

APA Style

Zhao, R. S., Chen, C., Van Horn, M., & Fogarty, N. D. (2026). Hierarchical YOLO-SAM: A Scalable Pipeline for Automated Segmentation and Morphometric Tracking of Coral Recruits in Time-Series Microscopy. Sensors, 26(8), 2291. https://doi.org/10.3390/s26082291

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop