Previous Article in Journal
Human Activity Recognition with Noise-Injected Time-Distributed AlexNet
Previous Article in Special Issue
Optimizing Maritime Search and Rescue Planning via Genetic Algorithms: Incorporating Civilian Vessel Collaboration
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

IDP-Head: An Interactive Dual-Perception Architecture for Organoid Detection in Mouse Microscopic Images

1
School of Software, Xinjiang University, Urumqi 830091, China
2
College of Electronic Engineering, Chengdu University of Information Technology, Chengdu 610225, China
3
The Xinjiang Technical Institute of Physics and Chemistry, Chinese Academy of Sciences, Urumqi 830011, China
4
School of Software Engineering, Chengdu University of Information Technology, Chengdu 610225, China
5
Key Laboratory of Remote Sensing Application and Innovation, Chongqing 401147, China
6
Dazhou Key Laboratory of Government Data Security, Sichuan University of Arts and Science, Dazhou 635000, China
*
Author to whom correspondence should be addressed.
Biomimetics 2025, 10(9), 614; https://doi.org/10.3390/biomimetics10090614
Submission received: 12 July 2025 / Revised: 8 September 2025 / Accepted: 9 September 2025 / Published: 11 September 2025
(This article belongs to the Special Issue Nature-Inspired Metaheuristic Optimization Algorithms 2025)

Abstract

The widespread application of organoids in disease modeling and drug development is significantly constrained by challenges in automated quantitative analysis. In bright-field microscopy images, organoids exhibit complex characteristics, including irregular morphology, blurred boundaries, and substantial scale variations, largely stemming from their dynamic self-organization that mimics in vivo tissue development. Existing convolutional neural network-based methods are limited by fixed receptive fields and insufficient modeling of inter-channel relationships, making them inadequate for detecting such evolving biological structures. To address these challenges, we propose a novel detection head, termed Interactive Dual-Perception Head (IDP-Head), inspired by hierarchical perception mechanisms in the biological visual cortex. Integrated into the RTMDet framework, IDP-Head comprises two bio-inspired components: a Large-Kernel Global Perception Module (LGPM) to capture global morphological dependencies, analogous to the wide receptive fields of cortical neurons, and a Progressive Channel Synergy Module (PCSM) that models inter-channel semantic collaboration, echoing the integrative processing of multi-channel stimuli in neural systems. Additionally, we construct a new organoid detection dataset to mitigate the scarcity of annotated data. Extensive experiments on both our dataset and public benchmarks demonstrate that IDP-Head achieves a 5-percentage-point improvement in mean Average Precision (mAP) over the baseline model, offering a biologically inspired and effective solution for high-fidelity organoid detection.

1. Introduction

The advent of three-dimensional (3D) organoid technology marks a paradigm shift in life sciences and regenerative medicine [1]. These stem cell-derived cultures, which faithfully recapitulate key structures and functions of their corresponding organs [2,3], offer unprecedented opportunities for research in areas such as disease modeling and high-throughput drug screening [4,5]. However, translating the immense potential of this technology into reliable experimental conclusions and novel biological findings is critically limited by the need for precise and efficient automated phenotypic quantification of vast amounts of microscopy images. This task is particularly challenging due to the inherent complexities of organoid images [6]: during dynamic cultivation, organoids exhibit high morphological heterogeneity [7], indistinct tissue boundaries, and substantial scale variations. Among these, scale variation stands out as one of the most prominent and driving challenges. Such proportional differences are not only reflected in the size disparities across different samples, but also manifest throughout the continuous deformation and growth of the same organoid during its developmental and dynamic evolutionary processes. These complex variations significantly compromise the stability of object detection and feature extraction, posing severe difficulties for existing visual recognition methods. Therefore, effectively addressing scale variation should be regarded as a core design motivation in the development of automated analysis systems, as it plays a decisive role in enhancing model robustness, generalization ability, and practical applicability.
Although deep learning, particularly Convolutional Neural Networks (CNNs), has become a mainstream technology for biomedical image analysis [8], existing object detectors—whether two-stage frameworks like Faster R-CNN [9] or one-stage ones like the YOLO [10] series—face two fundamental constraints when directly applied to organoid detection. The first lies in the limitation of spatial representation: the local computational paradigm of standard convolutional kernels inherently restricts their effective receptive field, making it difficult to capture the long-range spatial dependencies necessary for large, irregular organoids. The second is the bottleneck in modeling channel-wise semantics: while methods like the seminal SENet introduce channel attention, their reliance on linear fully-connected layers is insufficient for capturing the inherent higher-order, non-linear inter-channel correlations present in complex biological images. Departing from conventional detectors based on Convolutional Neural Networks (CNNs), the field of visual recognition, particularly in biomedical imaging, has witnessed the rise of novel architectures such as pure Transformers and hybrid convolutional-attention models. The key strength of these architectures is their self-attention mechanism, which empowers them to model long-range spatial dependencies and intricate inter-channel relationships, thereby overcoming the inherent locality bias of traditional convolutions. Pioneering technologies like Vision Transformer (ViT) and its hierarchical variant, Swin Transformer, exemplify the significant potential of this approach [11].
It is noteworthy that most current improvements in object detection focus on the network’s backbone or feature pyramid layers [12]. However, the internal feature enhancement design of the detection head—the pivotal component that translates high-level semantic features [13] into final predictions (classes and bounding boxes)—remains a largely overlooked research gap. To fill this critical gap, our work draws inspiration from the remarkable efficiency of the human visual system. We design a novel Interactive Dual-Perception Head (IDP-Head), founded on a biomimetic computational paradigm, and seamlessly integrate it into the RTMDet framework [14]. Its core idea is to reshape and enhance input features through two synergistic modules designed to emulate biological perception mechanisms before they are passed to the classification and regression branches [15]. We propose the Large-Kernel Global Perception Module (LGPM) and the Progressive Channel Synergy Module (PCSM) to enhance feature representation for organoid analysis. Inspired by the perceptual mechanism of the visual cortex for holistic scenes, the LGPM establishes long-range spatial dependencies to overcome the limited receptive fields of conventional convolutions, enabling the precise capture of macroscopic organoid structures. Concurrently, the PCSM emulates the brain’s approach to integrating multi-channel information. It models non-linear inter-channel relationships and adaptively enhances salient features via a learnable weighting parameter, thereby improving the model’s recognition accuracy for small-scale or low-contrast organoids [16].
The main contributions of this study are as follows:
We propose the LGPM module to address the challenge of complex organoid morphology. This module effectively captures long-range spatial dependencies, enabling the model to perceive the holistic structure of irregularly shaped organoids and significantly improving detection accuracy under high morphological variance.
We design the PCSM to achieve precise localization of organoids with indistinct boundaries. By adaptively modeling non-linear inter-channel relationships, the PCSM enhances crucial semantic features while suppressing background noise, thereby reducing ambiguity and improving detection accuracy in low-contrast bright-field images.
We construct a comprehensive, well-annotated organoid detection benchmark dataset. This provides reliable data support for quantitative studies of organoids and facilitates reproducible evaluations and method comparisons.
Experimental results on two datasets demonstrate that the proposed method performs competitively on key metrics such as mAP, underscoring its robustness and ability to generalize.
The remainder of this paper is organized as follows. Section 2 reviews the applications and developments of object detection techniques in the field of medical imaging. Section 3 elaborates on the experimental materials and our proposed methodology, with its core components being the IDP-Head and the Multi-branch Prediction module. Section 4 presents the overall experimental design, including the setup for comparative experiments and ablation studies, as well as the corresponding evaluation metrics and results. Section 5 provides an in-depth discussion and analysis of the experimental results, covering the interpretation of key metrics such as the confusion matrix, precision, and recall. Finally, Section 6 concludes this paper by summarizing our research work and provides an outlook on future research directions.

2. Related Work

This section reviews the relevant literature from two primary aspects. We first elucidate the performance bottlenecks of generic object detection methods when applied to the analysis of dense organoid images. Subsequently, by examining advanced visual perception techniques, we observe that the majority of prior research has focused on optimizing the network backbone, thereby highlighting a critical research gap in the design of the detection head.

2.1. Application of Object Detection in Medical Image Analysis

Early medical image analysis relied on conventional image processing techniques [17], such as thresholding and morphological operations [18]. However, these methods exhibit limited robustness and generalization capabilities when faced with the complexity and diversity of organoid images. Consequently, object detection [19] frameworks based on deep learning have become the mainstream paradigm in this field. Two-stage detectors, represented by Faster R-CNN, excel in tasks requiring high precision, like lesion detection, but their significant computational overhead limits their application in high-throughput screening scenarios. For this reason, one-stage detectors, such as the YOLO series and RTMDet, are favored for their exceptional efficiency. Nevertheless, when directly applied to dense organoid culture environments, these general-purpose frameworks often fail to achieve robust and accurate detection due to challenges like severe object overlap [20], dramatic scale variations, and blurred boundaries [21]. Fundamentally, their performance bottleneck lies in the fact that the design of their core feature extraction and perception modules is not fully adapted to the unique challenges of biological imagery.

2.2. Advanced Visual Perception Mechanisms and the Research Gap in Detection Head Design

To enhance feature representation, researchers have explored various advanced visual perception mechanisms from two primary dimensions: spatial and channel. In the spatial dimension, large-kernel convolution designs, represented by recent developments like ConvNeXt and RepLKNet [22], have demonstrated potential, surpassing traditional spatial attention. They can effectively expand the receptive field to capture long-range dependencies while maintaining computational efficiency. In the channel dimension, channel attention mechanisms [23], led by the seminal SENet [24], enhance key features by explicitly modeling inter-channel correlations. However, their simplified linear modeling approach struggles to capture complex non-linear semantics. Notably, a crucial point that is often overlooked is that most of these advanced perception mechanisms are applied to the network backbone [25]. Consequently, the detection head—the “last mile” responsible for translating high-level features into final predictions—remains deficient in designs for synergistic enhancement.

3. Materials and Methods

3.1. Datasets

To comprehensively evaluate our proposed method, we utilized two mouse organoid datasets with distinct characteristics and challenges: Mouse2Org and TelluOrg. The first, Mouse2Org, is a binary-class detection dataset containing 508 images [26], with its annotations converted from a semantic segmentation project; this conversion process is illustrated in Figure 1. In contrast, the TelluOrg dataset introduces a higher level of complexity, comprising 840 bright-field microscopy images that [27] encompass four distinct developmental stages of organoids, examples of which are shown in Figure 2. All the organoids in this dataset originated from crypts isolated from whole mouse small intestines. Based on clear morphological features, these organoids were manually classified into four categories: Cyst, defined as vesicular structures without any budding; Early Organoid, which has formed one to two crypts; Late Organoid, exhibiting three or more mature budding structures; and Spheroid, an independent spherical category with no apparent crypts. The TelluOrg dataset contains a total of 23,066 annotated instances. Together, these two datasets form a comprehensive evaluation benchmark, spanning from basic binary-class detection to complex, imbalanced, multi-class scenarios.

3.2. Description of Compared Models

To comprehensively validate the effectiveness and advancement of our proposed method, Org-IDPNet, we selected a diverse and representative range of mainstream object detectors for comparison, ensuring both breadth and depth in our evaluation. The compared methods span multiple paradigms: from the classic two-stage detector Faster R-CNN to balanced single-stage models such as TOOD and CenterNet; from Transformer-based architectures that are a current research hotspot, like Deformable-DETR and DINO, to the efficient lightweight designs represented by EfficientDet and YOLOv11; and even emerging detection paradigms, as exemplified by DiffusionDet. Furthermore, we conduct a direct comparison between our model and RTMDet-L. As our baseline model, this comparison most directly demonstrates the performance improvement brought by our proposed IDP-Head.

3.3. Evaluation Metrics

To provide a comprehensive and rigorous assessment of our model’s performance, we employed a suite of standard metrics widely recognized in the field of object detection. The evaluation hinges on the concepts of precision and recall, which are defined as follows:
Precision = TP TP + FP
Recall = TP TP + FN
where TP (True Positives), FP (False Positives), and FN (False Negatives) are determined based on the Intersection over Union (IoU) between the predicted bounding box ( B p ) and the ground-truth box ( B g t ). The IoU is calculated as follows:
IoU = area ( B p B g t ) area ( B p B g t )
The primary metric for accuracy is mean Average Precision (mAP). For a single object class, the Average Precision (AP) is computed by calculating the area under the precision–recall curve, which can be expressed as follows:
AP = 0 1 p ( r ) d r
where p ( r ) is the precision as a function of the recall r. The mAP is then the mean of AP values over all C object classes:
mAP = 1 C i = 1 C AP i
In line with standard benchmarks, our principal reported metric is the mean AP calculated over a range of IoU thresholds from 0.5 to 0.95 with a step size of 0.05 (often denoted as mAP [0.5:0.95]). Additionally, to provide a more nuanced view, we report the AP at a lenient threshold, mAP50 (IoU = 0.5), and at a strict threshold, mAP75 (IoU = 0.75), which better reflects the model’s precise localization capability. To further evaluate the model’s adaptability to object size variations, we analyze its performance on small (mAPS, medium (mAPM), and large (mAPL) objects, as defined by the COCO evaluation criteria.
Beyond accuracy, model efficiency is a key focus. We assess this using two metrics: the total number of trainable parameters (in millions, M) to measure the model’s spatial complexity, and the Floating Point Operations (FLOPs) (in GFLOPs) to quantify its computational complexity.

3.4. Implementation Details

All experiments were conducted on a Linux server equipped with eight NVIDIA RTX A6000 GPUs. The software environment was configured with Python 3.8, PyTorch 2.1.0, and MMDetection toolbox, accelerated by CUDA 11.8 and cuDNN 8.7. For our model, we employed the AdamW optimizer for 150 epochs with an initial learning rate of 0.006, a weight decay of 0.05, and a total batch size of 16 across all GPUs. For all other comparative models, the default hyperparameters from the MMDetection [28] toolbox were utilized without modification to ensure a fair comparison. The dataset was randomly partitioned into training, validation, and test sets using an 8:1:1 ratio.

3.5. Overview of Methods

We propose Org-IDPNet, a single-stage detector for high-throughput, high-precision organoid detection, whose architecture carefully balances accuracy and real-time performance (Figure 3). The framework comprises three key components: CSPDarkNet as the backbone for feature extraction, RTMFPN as the multi-scale feature fusion hub, and our custom IDP-Head tailored to organoid characteristics. In the processing pipeline, the input bright-field image first passes through CSPDarkNet, which efficiently generates a feature pyramid embedding rich, multi-level semantic information. This pyramid is then fed into RTMFPN for cross-scale alignment and fusion, ensuring the model simultaneously captures the macro-structures of large, mature organoids and the fine details of nascent, smaller organoids. The fused feature maps are finally processed by IDP-Head, where they undergo two parallel, complementary enhancement modules: LGPM, which captures long-range dependencies and global morphology in the spatial domain, and PCSM, which models inter-channel semantic relationships in a fine-grained manner. Outputs from these two modules are merged via element-wise addition to produce enriched feature representations combining both macro- and micro-level cues, which are then passed to parallel classification and regression branches to predict organoid classes and their bounding boxes.

3.6. Interactive Dual-Perception Head

We propose IDP-Head, an architecture built upon the RTMDet framework that simultaneously captures local details and global semantic dependencies for organoid object detection. Given a multi-scale feature pyramid { X i } i = 1 N , where each input feature map is denoted as X R B × C × H × W , our head consists of two parallel enhancement branches. The LGPM, which focuses on spatial structure modeling (as illustrated in Figure 4), constructs multi-scale representations from local to global by progressively enlarging the effective receptive field to capture both fine-grained textures and global contours, and by aggregating the resulting multi-scale information into spatial weights for adaptive feature modulation. In this way, the model simultaneously enhances the weak signals of small organoids while preserving the global consistency of large organoids or organoids undergoing growth and deformation. The PCSM, which models channel dependencies (as shown in Figure 5), is the parallel branch.
Their outputs are fused via element-wise addition to form a unified enhanced representation:
F = Z LGPM + Z PCSM
The LGPM is designed to simulate a large receptive field using efficient depthwise convolutions. It first applies two separable convolutions along horizontal and vertical directions:
U 1 = DWConv ( 1 , K ) ( X ) , U 2 = DWConv ( K , 1 ) ( U 1 ) ,
followed by dilated depthwise convolutions with the dilation rate r = 2 :
U 3 = DWConv ( 1 , K ) , r = 2 ( U 2 ) , U 4 = DWConv ( K , 1 ) , r = 2 ( U 3 ) .
Then, 1 × 1 convolution and SiLU activation are used to generate a spatial attention map:
A LGPM = σ Conv 1 × 1 ( U 4 ) ,
which is applied element-wise to the input:
Z LGPM = X A LGPM .
This path enables progressive spatial perception from local textures to global contours. Complementing this spatial enhancement, the PCSM is designed to model long-range dependencies across channels, a process it initiates with global average pooling:
X ^ = 1 H W h = 1 H w = 1 W X : , : , h , w ,
followed by group normalization:
X ^ norm = GroupNorm ( X ^ ) .
The tensors X, Y, and Z are generated using depthwise convolutions:
X = DWConv X ( X ^ norm ) , Y = DWConv Y ( X ^ norm ) , Z = DWConv Z ( X ^ norm ) .
X , Y , Z = Reshape ( X , Y , Z ; h , d ) .
After reshaping into h attention heads, scaled dot-product attention is computed as follows:
A 1 = Softmax X Y d , O = A 1 · Z .
The output is then reshaped back, averaged spatially, and activated to generate channel attention weights:
W PCSM = σ Mean h , w Rearrange 1 ( O ) ,
Z PCSM = X W PCSM .
Multi-branch Prediction
The fused feature F is passed into parallel classification and regression branches. The classification output is defined as
P cls R B × ( A · C ) × H × W ,
and the bounding box regression branch produces
P reg R B × ( A · 4 ) × H × W .
Here, B is the batch size, A is the number of anchors per location, C is the number of classes, and H and W are the spatial dimensions (height and width) of the feature map.
This interactive dual-pathway design enables IDP-Head to effectively combine structural context and channel-wise semantics, improving detection accuracy and robustness for morphologically complex organoid targets.

4. Experimental Results

This subsection details the comparative experiments and ablation studies conducted on the TelluOrg and MouseOrg datasets, and presents the experimental results based on key evaluation metrics.

4.1. Comparative Experiments on MouseOrg

To validate the performance of the proposed Org-IDPNet in a basic binary detection task, we compared it against several state-of-the-art object detectors. The quantitative results are summarized in Table 1. Org-IDPNet achieves the best performance across all key precision metrics, with an mAP of 76.7%, substantially outperforming all competing methods, including CO-DINO (74.5%) and YOLOv11 (74.3%). Under the more stringent localization criterion mAP75, our method also ranks first, demonstrating its superior bounding box regression capability. Crucially, compared to our baseline RTMDet-L, Org-IDPNet attains a net gain of +9.7 percentage points in mAP, while simultaneously reducing the model size and computational cost. This result convincingly confirms that the proposed IDP-Head can efficiently enhance the detector’s representational power for complex organoid features. Moreover, Org-IDPNet achieves the highest detection accuracy at small, medium, and large scales (mAPs/mAPm/mAPl), highlighting its strong multi-scale adaptability.
In addition to numerical comparisons, we further illustrate the practical detection performance of Org-IDPNet through visualizations in Figure 6. The figure juxtaposes representative bright-field images of organoids, corresponding ground-truth annotations, and our model’s predicted bounding boxes. As shown, Org-IDPNet accurately localizes the vast majority of organoids in the field of view, with high overlap between predictions and ground truth. Even for organoids exhibiting irregular shapes, blurred boundaries, or close proximity, our model reliably delineates precise contours, yielding minimal false negatives and false positives. These qualitative results provide direct evidence of the model’s robust capability for high-precision organoid detection in complex bright-field imagery.

4.2. Comparative Experiments on TelluOrg

In the second set of comparative experiments conducted on the TelluOrg dataset, we faced significant challenges due to its highly imbalanced class distribution and frequent overlaps among targets. These characteristics impose more stringent requirements on the robustness and generalization capability of detection models. As shown in Table 2, we performed a comprehensive evaluation of various state-of-the-art detectors on this dataset. The experimental results clearly demonstrate that our method consistently achieved the best performance across all critical accuracy metrics. Specifically, it attained a mean Average Precision (mAP) of 66.7%, markedly surpassing all compared approaches, including the strong baseline RTMDet-L (63.7%) and the advanced CO-DINO (60.6%). These findings further validate the outstanding balance of our method between accuracy and efficiency, as it achieves the highest precision while maintaining relatively low model complexity and computational overhead.
This study conducts a detailed comparison of Org-IDPNet against EfficientDet, DINO-4scale, and the baseline RTMDet-L model, evaluating per-category mean Average Precision (mAP) on organoid images (in Table 3). The proposed Org-IDPNet, as an advanced single-stage detector, consistently outperforms its counterparts: it surpasses the strong single-stage baseline RTMDet-L by 4.9 % and 4.6 % mAP on the “Spheroid” and “Organoid0” categories, respectively, and significantly exceeds the classic EfficientDet. Moreover, when compared with the Transformer-based DINO-4scale—whose architecture has exhibited excellent performance in general object detection—Org-IDPNet demonstrates even greater advantages on specialized organoid imagery, with DINO’s metrics falling substantially behind. These comprehensive, category-wise results convincingly demonstrate that Org-IDPNet is better suited than other leading single-stage and Transformer-based detectors for the complex task of organoid detection.
The quantitative advantages of our model are further corroborated by qualitative visualization results. As illustrated in Figure 7, we compared the prediction maps of Org-IDPNet against those of EfficientDet, DINO-4scale, and the baseline RTMDet-L on the same complex scene. It is clearly observable that EfficientDet suffers from numerous missed detections (false negatives), while DINO-4scale struggles with irregularly shaped organoids. The baseline model, RTMDet-L, exhibits imprecise localization for certain organoids with ambiguous boundaries. In contrast, our Org-IDPNet demonstrates superior detection performance. Not only does it successfully identify the majority of organoids in the field of view, including difficult samples missed by other models, but its generated bounding boxes also achieve the tightest fit to the actual object contours.

4.3. Ablation Study

Ablation studies were conducted on two datasets to dissect the individual contributions and synergistic effects of the proposed LGPM and PCSM modules. As shown in Table 4, integrating either module alone enhances the baseline performance, validating their independent effectiveness.
Further analysis reveals that the LGPM module significantly enhances the precision of bounding box regression through its localization-guided multi-scale modeling, an improvement especially prominent for the high-threshold metric mAP75, which is more sensitive to localization accuracy, yielding gains of +1.8% on the MouseOrg dataset and +1.4% on the TelluOrg dataset. In contrast, the PCSM module, by employing a channel-spatial progressive enhancement strategy, effectively boosts feature discriminability and recall; consequently, its advantages are more pronounced in terms of improvement in the overall mAP and the low-threshold metric mAP50, with recorded gains of +2.2% and +1.1% on MouseOrg, respectively. When used in combination, they exhibit a significant complementary effect, achieving performance growth across all evaluation metrics that surpasses the linear superposition of their individual contributions, thus demonstrating a synergistic “1 + 1 > 2” advantage. Mechanistically, this synergistic performance enhancement stems from LGPM’s effective suppression of localization error and PCSM’s significant mitigation of classification/miss error, providing a solid and interpretable theoretical foundation for the exceptional performance of the combined model.

5. Discussion

Beyond the aggregate mAP metric, we conducted a comprehensive and fine-grained analysis to better understand the model’s performance in the organoid detection task. This investigation moves beyond quantifying correct predictions to probe the specific nature and sources of error. To this end, three complementary indicators were jointly assessed: the confusion matrix, precision, and recall. The confusion matrix offers a qualitative perspective, visually revealing inter-class misclassifications and pinpointing weaknesses in distinguishing morphologically similar organoids. Concurrently, precision and recall provide quantitative insights by measuring the accuracy of positive predictions and the completeness of detections, respectively. By integrating these metrics, we achieve a holistic evaluation of the model’s behavior, elucidate the trade-off between its predictive accuracy and detection coverage, and arrive at a more robust judgment of its reliability for practical applications.

5.1. Confusion Matrix Analysis

As illustrated in Figure 8, we analyzed the normalized confusion matrix, where labels y1 through y5 represent organoid0, organoid1, organoid3, spheroid, and background, respectively. The results clearly demonstrate that the proposed Org-IDPNet exhibits a significant advantage in class-wise discrimination. It achieves the highest classification accuracy across several key categories—most notably y1 (69%) and y3 (72%)—among all evaluated models, while also exhibiting a superior ability to suppress inter-class confusion.
A prominent example of this strength is the model’s performance on background classification: unlike other models that frequently misclassify a large portion of background (y5) as foreground objects, Org-IDPNet restricts this false positive rate to just 32%, the lowest among all competitors. Furthermore, in the case of visually and structurally similar classes such as y1 and y2, Org-IDPNet demonstrates enhanced robustness and discrimination capacity.
These qualitative findings are well aligned with quantitative metrics, jointly confirming the outstanding performance of the Org-IDPNet architecture in improving organoid detection accuracy and reducing misclassification between classes.

5.2. Precision Analysis

The class-wise precision comparison, as shown in Table 5, further validates the superiority of Org-IDPNet. Our model consistently outperforms both the baseline RTMDet-L and the state-of-the-art detector TOOD across all categories, achieving notably higher precision scores. Most strikingly, Org-IDPNet significantly improves precision for challenging categories such as “spheroid”, increasing it from 0.351 (RTMDet-L) to 0.549—an impressive gain of nearly 20 percentage points. Similarly, for ”organoid0”, precision improves from 0.484 to 0.598. Compared to TOOD, Org-IDPNet also demonstrates a clear advantage, with a 7-point gain in “organoid3” (0.551 vs 0.483). These consistent improvements highlight Org-IDPNet’s stronger discriminative capability, especially in difficult cases, and underscore its ability to make more confident and accurate positive predictions while effectively reducing false positives.

5.3. Recall Analysis

In terms of recall performance, Org-IDPNet also demonstrates highly competitive results. As shown in Figure 9, our model achieves the highest recall among all compared methods for “organoid0” (0.68) and “organoid1” (0.56), effectively reducing the missed detection rate. Although the baseline model RTMDet-L achieves slightly higher recall in “organoid3” and “spheroid”, its performance on other categories—such as only 0.53 for organoid1—is suboptimal, indicating weaker overall consistency. In contrast, Org-IDPNet maintains high recall across all categories, highlighting its superior and balanced capability to retrieve true positives. Moreover, compared with Transformer-based architectures such as Ddq-DETR and DiffusionDet, as well as classical detectors like EfficientDet and TOOD, Org-IDPNet consistently achieves better recall. These results strongly suggest that the proposed model not only excels in feature extraction, but also enables more comprehensive and reliable detection of organoid targets within the image.

6. Conclusions

This study addresses the challenge of accurate detection in automated organoid analysis, which is often hindered by the morphological diversity, dense overlaps, and ambiguous boundaries of organoid structures. To this end, we propose Org-IDPNet, a novel and efficient single-stage detection framework specifically tailored for this task. At the core of Org-IDPNet is IDP-Head, which integrates two complementary modules: the LGPM captures rich multi-scale spatial structures and contextual information through parallel pathways, while the PCSM models long-range dependencies across feature channels. Together, these modules enhance the model’s representational and discriminative power.
Comprehensive experiments validate the effectiveness of our approach. First, ablation studies not only confirm the individual contributions of the LGPM and PCSM, but also highlight their strong synergistic effect, where their combination yields performance gains greater than the sum of their parts. Second, in extensive comparisons with leading detectors—including the strong baseline Rtmdet-L, the classical EfficientDet, and the Transformer-based Dino-4scale—Org-IDPNet achieves superior performance across all major evaluation metrics, including mAP, mAP50, and mAP75.
Further class-wise analyses reveal the model’s inherent advantages. Precision and recall metrics jointly demonstrate that Org-IDPNet effectively balances false positive suppression with minimized missed detections. Qualitative results, including visualizations and confusion matrix analysis, provide intuitive evidence that the model can reliably distinguish morphologically similar classes and accurately detect challenging samples that are often misclassified by other methods. Although the LGPM and PCSM strengthen multi-scale and channel-dependency modeling, the network can still miss or mis-detect targets in scenarios with very small objects, severe occlusion, or extreme crowding. When multiple organoids are highly adherent/closely connected, the model may also collapse them into a single instance. In addition, the current study focuses on detection only, and does not yet provide an end-to-end solution for fine-grained instance segmentation, morphological quantification, or longitudinal spatiotemporal analysis of organoid growth.
Looking ahead, we plan to extend this work in several directions. First, we aim to expand the current 2D detection framework [42] into a 3D paradigm, enabling more comprehensive segmentation of organoid structures and tracking of their dynamic growth processes. Second, we will explore transferring the core design of Org-IDPNet to broader biomedical imaging applications, such as nuclear detection in histopathological slides [43] or tumor region identification. Finally, efforts will be made toward model lightweighting and deployment optimization, facilitating seamless integration into existing microscopic imaging and analysis platforms, and ultimately advancing the automation of life science research.

Author Contributions

Conceptualization, Y.Y.; methodology, Y.Y., C.F. and P.W.; software, Y.Y. and P.W.; validation, Y.Y. and X.Z.; writing—review and editing, Y.Y. and C.F.; supervision, Y.Y. and C.F.; project administration, Y.Y. and C.F.; data curation, X.Z.; writing—original draft preparation, X.Z.; visualization, X.Z.; funding acquisition, X.Z.; writing—original draft preparation, P.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Sichuan Science and Technology Plan: Research and Application of Key Technologies of Command and Equipment for Hail Suppression in Xinjiang (2024YFHZ0151), Sichuan University students innovation and entrepreneurship training program (S202410621082), Chengdu University of Information Technology key project of education reform (JYJG2024206), the open project of Dazhou Key Laboratory of Government Data Security under Grant (ZSAQ202501, ZSAQ202502, ZSAQ202507), and the Key Laboratory of Remote Sensing Application and Innovation (LRSAI-2025004).

Data Availability Statement

Data requests can be emailed to the corresponding author’s email address.

Acknowledgments

Many thanks to the editorial board and reviewers for their efforts on our article!

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Lee, M.J.; Lee, J.; Ha, J.; Kim, G.; Kim, H.-J.; Lee, S.; Koo, B.-K.; Park, Y.K. Long-term three-dimensional high-resolution imaging of live unlabeled small intestinal organoids via low-coherence holotomography. Exp. Mol. Med. 2024, 56, 2162–2170. [Google Scholar] [CrossRef]
  2. Ewoldt, J.K.; DePalma, S.J.; Jewett, M.E.; Karakan, M.Ç.; Lin, Y.-M.; Mir Hashemian, P.; Gao, X.; Lou, L.; McLellan, M.A.; Tabares, J.; et al. Induced pluripotent stem cell-derived cardiomyocyte in vitro models: Benchmarking progress and ongoing challenges. Nat. Methods 2025, 22, 24–40. [Google Scholar] [CrossRef]
  3. Shoji, J.-Y.; Davis, R.P.; Mummery, C.L.; Krauss, S. Global literature analysis of organoid and organ-on-chip research. Adv. Healthc. Mater. 2024, 13, 2301067. [Google Scholar] [CrossRef] [PubMed]
  4. Rachmian, N.; Medina, S.; Cherqui, U.; Akiva, H.; Deitch, D.; Edilbi, D.; Croese, T.; Salame, T.M.; Ramos, J.M.P.; Cahalon, L.; et al. Identification of senescent, TREM2-expressing microglia in aging and Alzheimer’s disease model mouse brain. Nat. Neurosci. 2024, 27, 1116–1124. [Google Scholar] [CrossRef] [PubMed]
  5. Merrill, N.M.; Kaffenberger, S.D.; Bao, L.; Vandecan, N.; Goo, L.; Apfel, A.; Cheng, X.; Qin, Z.; Liu, C.-J.; Bankhead, A.; et al. Integrative drug screening and multiomic characterization of patient-derived bladder cancer organoids reveal novel molecular correlates of gemcitabine response. Eur. Urol. 2024, 86, 434–444. [Google Scholar] [CrossRef] [PubMed]
  6. Jain, A.; Gut, G.; Sanchis-Calleja, F.; Tschannen, R.; He, Z.; Luginbühl, N.; Zenk, F.; Chrisnandy, A.; Streib, S.; Harmel, C.; et al. Morphodynamics of human early brain organoid development. Nature 2025, 644, 1010–1019. [Google Scholar] [CrossRef]
  7. Papargyriou, A.; Najajreh, M.; Cook, D.P.; Maurer, C.H.; Bärthel, S.; Messal, H.A.; Ravichandran, S.K.; Richter, T.; Knolle, M.; Metzler, T.; et al. Heterogeneity-driven phenotypic plasticity and treatment response in branched-organoid models of pancreatic ductal adenocarcinoma. Nat. Biomed. Eng. 2024, 9, 836–864. [Google Scholar] [CrossRef]
  8. Saad, M.M.; O’Reilly, R.; Rehmani, M.H. A survey on training challenges in generative adversarial networks for biomedical image analysis. Artif. Intell. Rev. 2024, 57, 19. [Google Scholar] [CrossRef]
  9. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 39, 1137–1149. [Google Scholar] [CrossRef]
  10. Alif, M.A.R.; Hussain, M. YOLOv1 to YOLOv10: A comprehensive review of YOLO variants and their application in the agricultural domain. arXiv 2024, arXiv:2406.10139. [Google Scholar] [CrossRef]
  11. Wang, Y.; Deng, Y.; Zheng, Y.; Chattopadhyay, P.; Wang, L. Vision transformers for image classification: A comparative survey. Technologies 2025, 13, 32. [Google Scholar] [CrossRef]
  12. Quan, Y.; Zhang, D.; Zhang, L.; Tang, J. Centralized feature pyramid for object detection. IEEE Trans. Image Process. 2023, 32, 4341–4354. [Google Scholar] [CrossRef] [PubMed]
  13. Ji, S.; Zhang, H.; Zhang, J.; Fei, C.; Wang, X.; Liu, J.; Zhang, P. A three-stage model for infrared small target detection with spatial and semantic feature fusion. Expert Syst. Appl. 2026, 295, 128776. [Google Scholar] [CrossRef]
  14. Lyu, C.; Zhang, W.; Huang, H.; Zhou, Y.; Wang, Y.; Liu, Y.; Zhang, S.; Chen, K. Rtmdet: An empirical study of designing real-time object detectors. arXiv 2022, arXiv:2212.07784. [Google Scholar] [CrossRef]
  15. Xiao, T.; Yang, L.; Zhang, D.; Cui, T.; Zhang, X.; Deng, Y.; Li, H.; Wang, H. Early detection of nicosulfuron toxicity and physiological prediction in maize using multi-branch deep learning models and hyperspectral imaging. J. Hazard. Mater. 2024, 474, 134723. [Google Scholar] [CrossRef]
  16. Xu, C.; Zhang, T.; Zhang, D.; Zhang, D.; Han, J. Deep generative adversarial reinforcement learning for semi-supervised segmentation of low-contrast and small objects in medical images. IEEE Trans. Med. Imaging 2024, 43, 3072–3084. [Google Scholar] [CrossRef]
  17. Archana, R.; Jeevaraj, P.S.E. Deep learning models for digital image processing: A review. Artif. Intell. Rev. 2024, 57, 11. [Google Scholar] [CrossRef]
  18. Peng, L.; Lu, Z.; Lei, T.; Jiang, P. Dual-structure elements morphological filtering and local z-score normalization for infrared small target detection against heavy clouds. Remote Sens. 2024, 16, 2343. [Google Scholar] [CrossRef]
  19. Sun, Y.; Sun, Z.; Chen, W. The evolution of object detection methods. Eng. Appl. Artif. Intell. 2024, 133, 108458. [Google Scholar] [CrossRef]
  20. Wang, H.; Zhang, D.; Feng, J.; Cascone, L.; Nappi, M.; Wan, S. A multi-objective segmentation method for chest X-rays based on collaborative learning from multiple partially annotated datasets. Inf. Fusion 2024, 102, 102016. [Google Scholar] [CrossRef]
  21. Moglia, A.; Cavicchioli, M.; Mainardi, L.; Cerveri, P. Deep learning for pancreas segmentation on computed tomography: A systematic review. Artif. Intell. Rev. 2025, 58, 220. [Google Scholar] [CrossRef]
  22. Xue, F.; Pang, L.; Huang, X.; Ma, R. A lightweight convolutional network based on RepLKnet and Ghost convolution. In Proceedings of the 2024 4th International Conference on Electronic Information Engineering and Computer Communication (EIECC), Wuhan, China, 27–29 December 2024; pp. 1571–1576. [Google Scholar]
  23. Fu, C.; Du, B.; Zhang, L. Resc-net: Hyperspectral image classification based on attention-enhanced residual module and spatial-channel attention. IEEE Trans. Geosci. Remote Sens. 2024, 62, 1–15. [Google Scholar]
  24. Crist, W. Debunking the diffusion of senet. Board Game Stud. J. 2021, 15, 13–27. [Google Scholar] [CrossRef]
  25. Song, X.; Fang, X.; Meng, X.; Fang, X.; Lv, M.; Zhuo, Y. Real-time semantic segmentation network with an enhanced backbone based on Atrous spatial pyramid pooling module. Eng. Appl. Artif. Intell. 2024, 133, 107988. [Google Scholar] [CrossRef]
  26. Zhao, Z.; Chen, X.; Dowbaj, A.M.; Sljukic, A.; Bratlie, K.; Lin, L.; Fong, E.L.S.; Balachander, G.M.; Chen, Z.; Soragni, A.; et al. Organoids. Nat. Rev. Methods Prim. 2022, 2, 94. [Google Scholar] [CrossRef]
  27. Domènech-Moreno, E.; Brandt, A.; Lemmetyinen, T.T.; Wartiovaara, L.; Mäkelä, T.P.; Ollila, S. Tellu–an object-detector algorithm for automatic classification of intestinal organoids. Dis. Model. Mech. 2023, 16, dmm049756. [Google Scholar] [CrossRef]
  28. Chen, K.; Wang, J.; Pang, J.; Cao, Y.; Xiong, Y.; Li, X.; Sun, S.; Feng, W.; Liu, Z.; Xu, J.; et al. MMDetection: Open MMLab detection toolbox and benchmark. arXiv 2019, arXiv:1906.07155. [Google Scholar] [CrossRef]
  29. Chen, Q.; Wang, Y.; Yang, T.; Zhang, X.; Cheng, J.; Sun, J. You only look one-level feature. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 19–25 June 2021; pp. 13039–13048. [Google Scholar]
  30. Ahmed, A.; Tangri, P.; Panda, A.; Ramani, D.; Karmakar, S. Vfnet: A convolutional architecture for accent classification. In Proceedings of the 2019 IEEE 16th India Council International Conference (INDICON), Rajkot, India, 13–15 December 2019; pp. 1–4. [Google Scholar]
  31. Feng, C.; Zhong, Y.; Gao, Y.; Scott, M.R.; Huang, W. Tood: Task-aligned one-stage object detection. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, BC, Canada, 11–17 October 2021; pp. 3490–3499. [Google Scholar]
  32. Sun, P.; Zhang, R.; Jiang, Y.; Kong, T.; Xu, C.; Zhan, W.; Tomizuka, M.; Li, L.; Yuan, Z.; Wang, C.; et al. Sparse R-CNN: End-to-end object detection with learnable proposals. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 19–25 June 2021; pp. 14454–14463. [Google Scholar]
  33. Tan, M.; Pang, R.; Le, Q.V. EfficientDet: Scalable and efficient object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 10781–10790. [Google Scholar]
  34. Zhang, H.; Li, F.; Liu, S.; Zhang, L.; Su, H.; Zhu, J.; Ni, L.M.; Shum, H.-Y. Dino: Detr with improved denoising anchor boxes for end-to-end object detection. arXiv 2022, arXiv:2203.03605. [Google Scholar]
  35. Chen, S.; Sun, P.; Song, Y.; Luo, P. DiffusionDet: Diffusion model for object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 2–6 October 2023; pp. 19830–19843. [Google Scholar]
  36. Zhu, X.; Su, W.; Lu, L.; Li, B.; Wang, X.; Dai, J. Deformable DETR: Deformable transformers for end-to-end object detection. arXiv 2020, arXiv:2010.04159. [Google Scholar]
  37. Zhang, S.; Wang, X.; Wang, J.; Pang, J.; Lyu, C.; Zhang, W.; Luo, P.; Chen, K. Dense distinct query for end-to-end object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 7329–7338. [Google Scholar]
  38. Chen, Z.; Yang, C.; Chang, J.; Zhao, F.; Zha, Z.-J.; Wu, F. DDOD: Dive deeper into the disentanglement of object detector. IEEE Trans. Multimed. 2023, 26, 284–298. [Google Scholar] [CrossRef]
  39. Zong, Z.; Song, G.; Liu, Y. DETRs with collaborative hybrid assignments training. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 2–6 October 2023; pp. 6748–6758. [Google Scholar]
  40. Duan, K.; Bai, S.; Xie, L.; Qi, H.; Huang, Q.; Tian, Q. CenterNet: Keypoint triplets for object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 6569–6578. [Google Scholar]
  41. Khanam, R.; Hussain, M. YOLOv11: An overview of the key architectural enhancements. arXiv 2024, arXiv:2410.17725. [Google Scholar] [CrossRef]
  42. Wang, J.; Li, F.; He, L. A unified framework for adversarial patch attacks against visual 3D object detection in autonomous driving. IEEE Trans. Circuits Syst. Video Technol. 2025, 35, 4949–4962. [Google Scholar] [CrossRef]
  43. Zhou, H.; Chen, H.; Yu, B.; Pang, S.; Cong, X.; Cong, L. An end-to-end weakly supervised learning framework for cancer subtype classification using histopathological slides. Expert Syst. Appl. 2024, 237, 121379. [Google Scholar] [CrossRef]
Figure 1. Data Transformation Process of MouseOrg.
Figure 1. Data Transformation Process of MouseOrg.
Biomimetics 10 00614 g001
Figure 2. Categorical Overview of Tellorg Dataset.
Figure 2. Categorical Overview of Tellorg Dataset.
Biomimetics 10 00614 g002
Figure 3. Model Architecture.
Figure 3. Model Architecture.
Biomimetics 10 00614 g003
Figure 4. Structural Design of LGPM.
Figure 4. Structural Design of LGPM.
Biomimetics 10 00614 g004
Figure 5. Structural Design of PCSM.
Figure 5. Structural Design of PCSM.
Biomimetics 10 00614 g005
Figure 6. Comparison Between Ground Truth and Model Predictions.
Figure 6. Comparison Between Ground Truth and Model Predictions.
Biomimetics 10 00614 g006
Figure 7. Comparison of Detection Results Across Different Models.
Figure 7. Comparison of Detection Results Across Different Models.
Biomimetics 10 00614 g007
Figure 8. Confusion Matrices of Different Models, With Darker Colors Indicating Better Performance.
Figure 8. Confusion Matrices of Different Models, With Darker Colors Indicating Better Performance.
Biomimetics 10 00614 g008
Figure 9. Per-class Recall.
Figure 9. Per-class Recall.
Biomimetics 10 00614 g009
Table 1. Comparative Experimental Results on MouseOrg.
Table 1. Comparative Experimental Results on MouseOrg.
MethodsEvaluation Metrics
mAPmAP50mAP75mAP-smAP-mmAP-l ParametersFlops
Yolof [29]0.2850.6340.2090.1280.4050.40342.339 M61.222 G
Vfnet [30]0.3660.6690.3680.3590.3760.41032.709 M118 G
Tood [31]0.5380.8610.6070.4940.5850.52532.018 M123 G
Sparse-RCNN [32]0.1340.3020.1060.1170.1740.089106 M97.63 G
Faster-RCNN0.1790.4320.1040.1030.2670.12941.348 M134 G
EfficientDet [33]0.6070.8620.6710.4540.7170.6883.828 M2.303 G
DINO-4scale [34]0.4870.7230.6090.4550.5860.59047.54 M179 G
DiffusionDet [35]0.7040.8870.7840.6010.8000.743
Deformable-DETR [36]0.4250.7620.4370.3170.5190.47040.099 M123 G
DDQ-DETR-4scale [37]0.4120.7240.4570.3670.4980.426
Ddod [38]0.4730.8460.5040.4620.4980.42132.196 M111 G
Co-DETR-5scale [39]0.7450.9170.8570.6410.8530.81064.455 M
CenterNet [40]0.3910.7450.3670.4200.3840.32432.111 M123 G
YOLO V11 [41]0.7430.9340.8510.6180.8380.7962.58 M6.3 G
RTMDet-L0.6700.8940.7850.5570.7670.76552.255 M79.951 G
Org-IDPNet (Ours)0.7670.9410.8740.6380.8550.8495.548 M8.112 G
Table 2. Comparative Experimental Results on TelluOrg.
Table 2. Comparative Experimental Results on TelluOrg.
MethodsEvaluation Metrics
mAPmAP50mAP75mAP-smAP-mmAP-lParametersFLOPs
YOLOF0.1240.2670.0990.0090.0870.15142.409 M83.32 G
VFNet0.1050.2140.0910.0280.0870.08332.716 M161 G
TOOD0.3940.6050.4410.1530.4050.34632.025 M168 G
Sparse-RCNN0.1060.2200.0910.0340.0900.090106 M130 G
Faster-RCNN0.1090.2560.0740.0390.0950.13741.364 M178 G
EfficientDet0.4790.7030.5340.1310.4710.6743.83 M2.313 G
DINO-4scale0.3670.5300.4250.1830.3740.43547.546 M235 G
DiffusionDet0.4890.6990.5530.1840.5010.494
Deformable-DETR0.3220.5660.3250.1100.3120.44940.099 M165 G
DDQ-DETR-4scale0.5190.7180.5990.1960.5300.590
DDOD0.2550.4320.2800.0990.2320.22932.203 M151 G
Co-DINO-5scale0.6060.7880.7040.2580.6150.66664.483 M
CenterNet0.2020.3760.1970.0770.2060.17032.118 M167 G
YOLO V110.5950.7850.6890.2240.6100.5922.583 M6.3 G
RTMDet-L0.6370.8330.7370.2530.6450.65952.257 M79.958 G
Org-IDPNet (Ours)0.6670.8600.7750.2550.6840.7365.549 M8.115 G
Table 3. Comparison Between EfficientDet, DINO-4scale, RTMDet-L, and Org-IDPNet on Different Categories.
Table 3. Comparison Between EfficientDet, DINO-4scale, RTMDet-L, and Org-IDPNet on Different Categories.
EfficientDetDINO-4scale
CategorymAPmAP50mAP75mAPmAP50mAP75
Organoid00.3600.6320.3580.3940.5890.476
Organoid10.4050.6120.4510.3070.4460.355
Organoid30.6010.8470.7050.4050.5980.428
Spheroid0.5490.7220.6240.3700.4860.438
RTmdet-LOrg-IDPNet (Ours)
CategorymAPmAP50mAP75mAPmAP50mAP75
Organoid00.5660.8110.6550.6120.8470.716
Organoid10.6270.8270.7250.6440.8470.744
Organoid30.7020.8800.8060.7140.8950.833
Spheroid0.6510.8150.7620.7000.8520.806
Table 4. Ablation Studies on Different Datasets.
Table 4. Ablation Studies on Different Datasets.
MouseOrgTelluOrg
MethodmAPmAP50mAP75mAPmAP50mAP75
base0.7340.920.8470.6420.8360.744
base + LGPM0.7530.9330.8650.6550.850.758
base + PCSM0.7560.9310.8620.650.8460.762
base + LGPM + PCSM0.7670.9410.8740.6670.860.775
Table 5. Per-class Precision.
Table 5. Per-class Precision.
Methodorganoid0organoid1organoid3Spheroid
Tood0.5050.4000.4830.593
RTMDet-L0.4840.4580.4560.351
Org-IDPNet0.5980.5070.5510.549
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yang, Y.; Fan, C.; Zhou, X.; Wei, P. IDP-Head: An Interactive Dual-Perception Architecture for Organoid Detection in Mouse Microscopic Images. Biomimetics 2025, 10, 614. https://doi.org/10.3390/biomimetics10090614

AMA Style

Yang Y, Fan C, Zhou X, Wei P. IDP-Head: An Interactive Dual-Perception Architecture for Organoid Detection in Mouse Microscopic Images. Biomimetics. 2025; 10(9):614. https://doi.org/10.3390/biomimetics10090614

Chicago/Turabian Style

Yang, Yuhang, Changyuan Fan, Xi Zhou, and Peiyang Wei. 2025. "IDP-Head: An Interactive Dual-Perception Architecture for Organoid Detection in Mouse Microscopic Images" Biomimetics 10, no. 9: 614. https://doi.org/10.3390/biomimetics10090614

APA Style

Yang, Y., Fan, C., Zhou, X., & Wei, P. (2025). IDP-Head: An Interactive Dual-Perception Architecture for Organoid Detection in Mouse Microscopic Images. Biomimetics, 10(9), 614. https://doi.org/10.3390/biomimetics10090614

Article Metrics

Back to TopTop