Next Article in Journal
An Experimental 10-Port Microwave System for Brain Stroke Diagnosis—Potentials and Limitations
Previous Article in Journal
Anomaly Detection in Nuclear Power Production Based on Neural Normal Stochastic Process
Previous Article in Special Issue
Optimized Machine Learning for the Early Detection of Polycystic Ovary Syndrome in Women
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Cross-Scale Hypergraph Neural Networks with Inter–Intra Constraints for Mitosis Detection

1
School of Artificial Intelligence and Computer Science, Nantong University, Nantong 226019, China
2
School of Information Science and Technology, Nantong University, Nantong 226019, China
3
Xinglin College, Nantong University, Nantong 226019, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(14), 4359; https://doi.org/10.3390/s25144359 (registering DOI)
Submission received: 12 May 2025 / Revised: 9 July 2025 / Accepted: 10 July 2025 / Published: 12 July 2025
(This article belongs to the Special Issue Recent Advances in Biomedical Imaging Sensors and Processing)

Abstract

Mitotic figures in tumor tissues are an important criterion for diagnosing malignant lesions, and physicians often search for the presence of mitosis in whole slide imaging (WSI). However, prolonged visual inspection by doctors may increase the likelihood of human error. With the advancement of deep learning, AI-based automatic cytopathological diagnosis has been increasingly applied in clinical settings. Nevertheless, existing diagnostic models often suffer from high computational costs and suboptimal detection accuracy. More importantly, when assessing cellular abnormalities, doctors frequently compare target cells with their surrounding cells—an aspect that current models fail to capture due to their lack of intercellular information modeling, leading to the loss of critical medical insights. To address these limitations, we conducted an in-depth analysis of existing models and propose an Inter–Intra Hypergraph Neural Network (II-HGNN). Our model introduces a block-based feature extraction mechanism to efficiently capture deep representations. Additionally, we leverage hypergraph convolutional networks to process both intracellular and intercellular information, leading to more precise diagnostic outcomes. We evaluate our model on publicly available datasets under varying imaging conditions, and experimental results demonstrate that our approach consistently outperforms baseline models in terms of accuracy.

1. Introduction

Breast cancer is the most commonly diagnosed cancer among women worldwide, accounting for 24.2% of all new cancer cases and 15% of cancer-related deaths in women [1,2]. Early detection of breast cancer significantly improves prevention and treatment outcomes. Currently, the most common and effective method for early screening is histopathological diagnosis using tissue sections. Among various histopathological features, mitotic activity is a key indicator of tumor aggressiveness [3,4]. Traditionally, pathologists manually examine hematoxylin and eosin (H&E)-stained slides to identify mitotic figures in gigapixel whole slide images (WSIs). This process is time-consuming and labor-intensive. Due to the global shortage of trained pathologists, manual diagnosis is often slow and costly.
As deep learning technology improves by leaps [5], AI models for processing pathological images have made significant breakthroughs, offering new perspectives for mitosis detection [6]. As illustrated in Figure 1, the current pipeline for automated mitosis detection typically involves several key steps. First, whole slide images are acquired and preprocessed. Then, expert pathologists annotate the mitotic figures. Common mitotic lesions in different tumors include breast carcinoma, neuroendocrine tumor, lymphosarcoma, and soft tissue sarcoma. Subsequently, the WSIs are divided into hundreds of tiled image patches, and neural networks are employed to detect mitotic cells within these patches.
In recent years, numerous deep learning techniques have been developed for mitosis detection in different types of tumors. These methods can be categorized into two primary groups: the first group utilizes object detection networks that are effective on natural images, such as Faster R-CNN [7] and RetinaNet [8]. However, these approaches often struggle with performance and lack interpretability, making them unsuitable for clinical use. The second group includes enhanced detection techniques that incorporate specific knowledge about tumors. D. Cireşan et al. [9] proposed a simple convolutional neural network (CNN) for mitosis detection and won first place in the 2012 ICPR Mitosis Detection Challenge. He was among the first to apply deep learning techniques to the task of mitosis detection. H. Chen et al. [10] designed a deep cascade network that first performs coarse localization of candidate cells, followed by a fine-grained classification model that incorporates knowledge from cross domains. Alom et al. [11] introduced an integrated multi-block reference scheme along with a novel confidence analysis strategy to improve overall detection performance. Although these methods have achieved significant performance, they still face several critical challenges:
  • Limited training data: In pathological slides, mitotic cells are significantly fewer than normal cells, leading to a severe class imbalance between positive and negative samples in the dataset. Furthermore, since annotation requires expert pathologists, the amount of data available for training is extremely limited. In addition, the structure and morphology of cells vary greatly, and an imbalanced distribution among different types of mitotic figures is very common in datasets. As a result, trained detection models often suffer from high false negative or false positive rates. Moreover, with the increasing complexity and parameter size of current benchmark models, overfitting on mitosis detection tasks has become a frequent issue.
  • Subtle features: Mitosis is a complex biological process and pathologists typically rely on nuclear morphology to determine whether a cell is undergoing mitosis. In H&E-stained slides, mitotic nuclei appear as dark blue dots, which are often difficult to distinguish from the background and can easily be confused with apoptotic cells that also appear as dark blue dots. In addition, the morphological differences between various cell types in tissue sections are minimal, and mitotic cells often resemble normal cells in appearance. These factors make the accurate detection of mitotic figures extremely challenging.
  • Neglecting cell relationships: In pathological diagnosis, valuable information lies not only in individual cellular features but also in the relationships between surrounding cells. For cells with ambiguous or indistinct features, pathologists often rely on comparisons with neighboring cells to determine whether mitosis is occurring. However, most current models lack the ability to effectively model intercellular context, resulting in the loss of critical diagnostic information.
In addition, we surveyed the recent top-performing methods from the MIDOG2022 challenge. The winning solution, Stain-Robust Mitotic Figure Detection for the Mitosis Domain Generalization Challenge [12], formulated mitosis detection as a pixel-level segmentation task with stain-invariant masks, while the runner-up, Sk-Unet Model with Fourier Domain for Mitosis Detection [13], employed frequency-domain adaptation to address domain shift. Although both methods achieved strong accuracy, they rely on labor-intensive pixel annotations and do not explicitly model the clinical prior knowledge or intercellular context used by pathologists. These limitations highlight ongoing challenges in bridging algorithmic performance with practical diagnostic needs.
Nowadays, Graph Neural Networks (GNNs) are extensively utilized in multiple domains [14,15,16,17], leading to the development of various graph-based enhancements. One notable advancement is the hypergraph convolutional network (HGNN) introduced by Feng et al. [18], which is designed to manage complex data. Unlike conventional graph networks, HGNN enables a hyperedge to link several vertices, providing greater flexibility in data modeling. Building on this concept, we created an Inter–Intra Hypergraph Neural Network that achieves improved accuracy in mitosis detection. In conclusion, the key contributions of this paper are as follows:
  • Design a Block-Based Mixed Mechanism (BBMM), using parallel convolutional modules to efficiently extract deep information and enrich the gradient flow during training. In the feature fusion phase of the model, use a Bottom–Up mechanism to recover non-abstracted spatial details. These operations enable efficient feature analysis at the inter-level.
  • Thoroughly analyze the shortcomings of existing mitosis detection models and apply the HGNN concept to the cell detection domain, effectively modeling the relationships between individual cells and cell populations. A novel hyperedge convolutional construction is designed to capture the visual features of different cells. These operations facilitate efficient feature analysis at the intra-level.
  • Test the model on a public dataset with multiple tumor labels and originating from different staining imaging conditions, achieving favorable results. According to this research, this is the first time the HGNN concept has been applied in the mitosis detection domain.

2. Materials and Methods

2.1. Overview

The YOLO series [19,20,21,22,23,24,25,26,27,28,29,30,31] of single-stage object detection networks have been widely applied in natural image processing due to their accuracy and real-time performance. The best-performing model in this series is currently YOLO11 [32]. The architecture of the YOLO11 network can be divided into three components: the backbone network, which extracts features from images; the neck network, which fuses multi-scale features; and the head network, which generates local candidate boxes. Although the original YOLO11 algorithm demonstrates strong generalization capabilities, it has some limitations. It has difficulty detecting small objects and does not effectively model visual relationships at the group level [33]. These limitations render the original model inadequate for tasks that involve mitosis detection.
In order to address the limitations of the baseline model in mitosis detection, we propose an enhanced network architecture, which is depicted in Figure 2. Specifically, we introduce a Block-Based Mixed Mechanism (BBMM) within the backbone of the network. This mechanism employs a parallel convolution strategy which is designed to enrich the gradient flow during the training process. As a result, it significantly improves the feature representation at the inter-level, allowing for more robust and detailed extraction of cell features. In the neck of the network, we integrate hypergraph convolution networks. These networks are capable of effectively modeling the relationships between individual cells, thereby facilitating the fusion of heterogeneous visual features at the intra-level. This integration ensures that the network can capture both the individual characteristics of cells as well as their interactions with neighboring cells. Furthermore, in the head of the network, we incorporate a Bottom–Up strategy. This strategy enhances the network’s capability to process cross-hierarchical information, enabling it to better understand the context and structure of cell populations. By jointly leveraging the individual-level cellular attributes and the population-level contextual information, our proposed design produces a high-precision mitosis detection framework. This framework not only improves the representational capacity of the network, but also enhances its structural awareness, making it more effective in mitosis detection with high accuracy and reliability.

2.2. Block-Based Mixed Mechanism (BBMM)

As illustrated in Figure 3, the C3K2 module mainly relies on a single module for convolutional operations aimed at feature extraction. Although it supports both feature extraction and fusion, its capacity to capture a variety of information flows is somewhat restricted. Furthermore, when addressing multi-scale features, the C3K2 module primarily utilizes a basic fusion approach, which limits its effectiveness in integrating features across different levels. These shortcomings greatly hinder the model’s ability to handle intricate cellular visual information, making the original YOLO11 architecture insufficient for accurate cell detection tasks.
Given the limited feature extraction capability of the C3K2 module, we attribute this limitation to its reliance on a single convolutional mechanism. To address this, we design the Block-Based Mixed Mechanism (BBMM). While retaining the original C3K2 module, BBMM incorporates two additional classic convolutional operations: 1 × 1 convolution and deformable convolution [34], as illustrated in Figure 4. This enhancement aims to enrich the representation of features by introducing both fine-grained channel transformations and adaptive spatial modeling.
The 1 × 1 convolution does not capture local spatial relationships within the input data but instead focuses on inter-channel interactions, allowing for channel-wise feature recalibration. To address the morphological variations of cells at different pathological stages, we propose using deformable convolution, which dynamically adjusts the receptive field to accommodate nonrigid deformations. By adopting a parallel strategy that integrates these three types of convolutions, we enhance the diversity and richness of the gradient flow during training, thus laying a stronger foundation for subsequent modeling in the neck module.

2.3. Hypergraph Neural Network

Our model refinement focuses primarily on the neck module, where we incorporate the concept of hypergraph convolution networks. Unlike traditional graph structures, where an edge connects only two vertices, a hypergraph allows a single hyperedge to connect multiple vertices simultaneously. The formal definition of a hypergraph is typically given as
G = ( V , E )
where V signifies the collection of vertices and E indicates the collection of hyperedges. In our approach, we decompose the grid-based visual features to create the vertex set V of a hypergraph. More specifically, the features obtained from five successive convolutional operations are combined through channel-wise concatenation to generate a hybrid feature representation X m . For each feature map with dimensions ( C i , H i , W i ) , where C i , H i and W i are its channel, height, and width, respectively, the resulting dimension of the hybrid feature X m is
C m = i = 1 5 C i .
where the upper limit 5 corresponds to the 5 successive convolutional outputs F 1 F 5 obtained from stages P3–P7 of YOLO11.
In a hypergraph, each feature point can be regarded as a vertex, which means that the total number of vertices corresponds to the total number of pixels on the feature map. With such a high number of vertices, a major challenge is to create an effective strategy for constructing hyperedges that can efficiently identify meaningful relationships between them. The construction method of the hypergraph can be visualized as shown in Figure 5.
We create the collection of hypergraph vertices denoted as V . To represent the neighborhood connections within the hypergraph module, we establish the set of hyperedges E using a distance threshold λ . For each feature point x u , we identify all feature points that are within a distance of less than λ and connect them to form a hyperedge with x v . A hyperedge e can be expressed as
e = { u | | x u x v | | 2 < λ , u V } ,
The Euclidean Norm · 2 encompasses all hyperedges that make up the hyperedge set E. The incidence matrix H for the hypergraph G = ( V , E ) is defined as follows:
H v e = 1 , if v e 0 , if v e
In hypergraph convolutional networks, increasing the number of layers can lead to a gradual loss of information during propagation. To mitigate this issue, we incorporate residual connections that allow the model to transmit input information directly to the output, thereby reducing the risk of information loss. To facilitate the propagation of high-order information within the hypergraph structure, we employ spatial hypergraph convolution along with residual connections. The computation process is outlined as follows:
X e = 1 | N v ( e ) | v N v ( e ) X v Θ e , X v = X v + 1 | N e ( v ) | v N e ( v ) X e
Let N v ( e ) denote the neighborhood of vertices associated with the hyperedge e, and N e ( v ) denote the neighborhood of hyperedges linked to the vertex v. Additionally, Θ e is a parameter that can be trained. With a feature matrix X for the vertices and an adjacency matrix H for the hypergraph, and assuming D v and D e are the degree matrices for the vertices and hyperedges, respectively, the hypergraph convolution can be formulated as follows:
H y p e r C o n v ( X , H ) = X + D v 1 H D e 1 H T X Θ
The expression D v 1 H D e 1 computes the normalized adjacency matrix, while H is the adjacency matrix of the graph. D v 1 H D e 1 H T X aggregates vertex features via hyperedges to identify higher-order connections among vertices. The matrix Θ is a learnable parameter that transforms the aggregated features, improving the expressiveness of the model. Additionally, a residual connection of X is included to maintain the original information about the feature and avoid any loss of information.
In summary, the hypergraph neural network addresses the shortcomings of conventional grid structures, improving the depth and detail of feature representations. This approach effectively merges individual-level inter-correlation features of cells with population-level intra-discriminative features, allowing for a successful integration of diverse features.

2.4. Bottom–Up

After defining the vertex set V and the hyperedge set E , we proceed to build the hypergraph convolution network. The output of this network encompasses higher-order, cross-level information. To better align with the head and improve detection capabilities, we implement a Bottom–Up mechanism [35] between the neck and head. As features move through deeper layers, spatial details tend to diminish due to the more abstract nature of the feature representations. As illustrated in Figure 2, the information processed by the hypergraph is integrated with the last three layers of the backbone’s output. This integration allows for the transfer of high-resolution details from the shallower layers to the deeper layers, thereby maintaining essential structural information and offsetting the loss of spatial details in the baseline.

3. Experiments and Results

3.1. Datasets

The evaluation dataset used in our experiments is the MIDOG2022 dataset [36] provided by the MICCAI Challenge. This dataset is specifically designed for detecting mitosis in different types of tumor cells. The MIDOG dataset contains 150 whole slide images (WSIs) annotated with various types of mitosis. These WSIs were captured using three different imaging conditions with scanners, each providing 50 WSIs. This necessitates higher robustness in the detection models due to the variability in imaging conditions. To facilitate network input, we performed preprocessing operations on all the images, dividing each complete WSI into patches of 640 × 640 pixels and then reconstructing the lesion annotations for each patch.
In total, the MIDOG2022 dataset includes 9501 annotated mitotic figures and 11,051 hard negative samples, ensuring a rich yet challenging training environment. Each WSI belongs to one of several tumor types, including breast carcinoma, lymphoma, lung carcinoma, melanoma, mast cell tumor, and neuroendocrine tumor, drawn from both human and canine subjects. These categories are approximately balanced within the mitotic class, but the overall dataset remains highly imbalanced, with mitotic figures constituting less than 10% of all annotated cells. To prevent data leakage, the dataset is split at the slide level into training (70%, 105 WSIs), validation (15%, 22 WSIs), and testing (15%, 23 WSIs) sets. After preprocessing, this resulted in approximately 8000 mitosis-positive patches and over 100,000 background patches. The substantial class imbalance and diversity in both tissue origin and imaging conditions make MIDOG2022 an ideal benchmark for evaluating model robustness and generalization.
It is worth noting that the specific cancer types corresponding to each WSI have not been disclosed by the dataset organizers, in order to avoid introducing potential bias. As a result, no tumor-specific stratification was applied during training or evaluation. Nevertheless, the dataset includes mitoses sampled from a variety of tumor tissues and scanning conditions, ensuring broad diversity in visual patterns and clinical relevance.

3.2. Implementation Details

The implementation of our method is based on the PyTorch framework (version = 1.8.0) and runs on an NVIDIA RTX 3090 GPU (Leadtek Research Inc., Shanghai, China). Our model is trained using the YOLO11n pre-trained model with the Adam optimizer [37]. The comparative experiments are conducted using MMDetection [38] and evaluated with COCO-style [39] metrics for quantitative analysis. These metrics include P (Precision), R (Recall), and mAP (Mean Average Precision). The specific calculation formulas for these metrics are as follows:
Precision = TP TP + FP
Recall = TP TP + FN
AP = 0 1 P ( R ) d R
mAP = 1 C i = 1 C AP i
ACC = TP + TN TP + FP + FN + TN
F 1 - score = 2 · p r e c i s i o n · r e c a l l p r e c i s i o n + r e c a l l
True positive (TP) refers to the count of instances that the model correctly classifies as positive. False positive (FP) is the number of negative instances that are incorrectly identified as positive by the model. False negative (FN) represents the instances where the model incorrectly classifies a positive instance as negative. True negative (TN), conversely, denotes the number of negative instances that are correctly identified as negative. Precision is derived by dividing TP by the total of TP and FP, offering a measure of the model’s accuracy in predicting positive outcomes. Recall is the ratio of TP to the sum of TP and FN, indicating the model’s capability in detecting positive instances [40]. Average Precision (AP) is the area under the Precision–Recall (P-R) curve [41], which can be computed by integrating the curve. The constant C indicates the number of categories. In multi-class classification, AP is computed for each class individually, and the average of these AP values yields the mean Average Precision (mAP), which accounts for the specific P-R curve of each class. Overall Accuracy (ACC) measures the proportion of all correctly classified samples—both positive (TP) and negative (TN)—to the total number of samples, providing a global indicator of recognition performance. Finally, the F1-score is the harmonic mean of Precision and Recall, balancing the trade-off between FP and FN to summarize a model’s effectiveness when both error types are of comparable concern.

3.3. Comparative Analysis

To ensure a comprehensive evaluation, we compare our method with several representative CNN-based detectors, including Faster R-CNN, RetinaNet, YOLOX, YOLO11, Sparse R-CNN, Cascade R-CNN, and RT-DETR. These models span classical two-stage frameworks, efficient single-stage detectors, and transformer-based architectures. Faster R-CNN and Cascade R-CNN are well-established two-stage models with strong localization performance. RetinaNet and YOLOX are popular one-stage detectors, with RetinaNet addressing class imbalance via Focal Loss. RT-DETR leverages transformer-based global attention for enhanced accuracy, while YOLO11 represents the latest YOLO variant balancing speed and precision. Collectively, these baselines form a diverse and rigorous benchmark for validating our approach. We strictly adhere to the rules for quantitative analysis, and the results of our tests on the same GPU are presented in Table 1. Here, mAP50 denotes the mean Average Precision computed at an IoU threshold of 0.50, Paramets (M) indicates the number of trainable parameters expressed in millions, and GLOPs (G) refers to the computational cost in giga floating-point operations for a single forward pass. Please note that the YOLO11 version used here is YOLOv11-l.
It can be observed that on the MIDOG2022 dataset, our proposed detection model consistently outperforms other object detection models. In particular, our model achieves a 6.4% improvement in mAP50 compared to the existing best model, which is the result of more accurate localization and classification by our model. Furthermore, our model improves in Precision and Recall by 2.8% and 5.2%, respectively. It is worth noting that all methods are trained solely on the official MIDOG2022 training set, which contains 150 whole slide images (WSIs) with about 9000 annotated mitotic figures—considerably smaller than pathological benchmarks such as CAMELYON17 [46], which comprises 1000 WSIs with tumor-region annotations, or DeepLesion [47], which includes more than 32,000 CT lesions. Hence, the superior results indicate that our model can maintain high detection accuracy even with limited training data, making it particularly suitable for fields such as mitosis detection, where data acquisition is challenging.
To further visualize the classification effectiveness, we compare the confusion matrices of the baseline model and our proposed approach. These matrices reveal how each model distinguishes between mitotic figures, hard negatives, and background, providing complementary insight to the numerical metrics reported earlier.
The confusion matrices in Figure 6 offer an intuitive comparison of the baseline model and our proposed approach. In contrast to the baseline, our model evidently demonstrates improved classification accuracy. Specifically, the number of correctly identified mitotic figures increases by 44 instances, while false positives related to background and hard negatives are significantly reduced. This improvement indicates that our model not only enhances mitosis detection sensitivity but also suppresses over-detection of non-mitotic structures. The reduction in misclassification between morphologically similar classes further suggests better inter-class separability. These results align with the Precision and Recall gains reported in Table 1, underscoring our model’s robustness in handling visually ambiguous samples.

3.4. Ablation Study

To thoroughly assess the effectiveness of each component in our proposed network architecture, ablation studies were conducted using the MIDOG2022 dataset. As illustrated in Table 2, our method consistently demonstrates substantial improvements across various evaluation metrics when compared to the baseline YOLO11-l model.
Specifically, after the introduction of the Block-Based Mixed Mechanism (BBMM) module, the model exhibits notable enhancements across all metrics. This clearly indicates that the parallel multi-convolutional structure within the BBMM module significantly boosts the model’s ability to capture and model complex cellular structures, thereby enriching the feature representation at the inter-level. Subsequently, upon incorporating the Hypergraph Convolutional Neural Network (HGNN) module, the model achieves remarkable improvements in both Average Precision (AP) and Average Recall (AR). This finding underscores the critical importance of modeling the intercellular structural relationships for the accurate recognition of high-heterogeneity lesions. By effectively capturing the interactions between cells, the HGNN module further enhances the model’s intra-level feature fusion capabilities. Furthermore, when both the BBMM and HGNN modules are integrated into the network, the system’s overall performance reaches its peak. The mean Average Precision (mAP) increases from 87.5 to 93.6, highlighting the synergistic benefits of combining inter-level and intra-level modeling. This substantial improvement demonstrates that the joint optimization of individual cell features and population-level contextual information significantly boosts the model’s detection accuracy and robustness.
It is worth noting that adjusting the detection threshold is crucial for performance optimization. In neural networks, performance related to information propagation typically follows a convex curve, indicating the existence of an optimal threshold. Empirical analysis can help determine a reasonable range for the threshold, while ablation studies can assist in identifying the precise value. The specific experimental results are shown in Figure 7.
Based on the experimental results, the performance of the proposed network follows a convex curve. As the threshold increases, the detection performance increases rapidly and remains relatively stable with only minor fluctuations in the threshold range of 7 to 10. We attribute this increase to the fact that lower thresholds result in reduced hypergraph connectivity, which does not fully take advantage of the high-order relationships among features. Conversely, higher thresholds connect more hypergraphs, leading to excessive information sharing and over-smoothing of features. This is evident in the performance drop observed when the threshold is set to 9, as shown in the figure. Therefore, our detection network is constructed using a distance threshold of 8.

3.5. Visualization Map

To intuitively demonstrate the accuracy of our model, we selected several representative comparison samples, as shown in Figure 8. Please note that due to the large image sizes, we cropped local regions for better visualization. The number inside each bounding box denotes the confidence score (0–1) output by the corresponding detector for the “mitotic figure” class; the larger the value, the more certain the model is about that prediction. Specifically, we have drawn the following two conclusions: (1) First, among the existing object detection models, the RT-DETR model generally outperforms the others, demonstrating the powerful capabilities of Transformer-based models. (2) Second, our model achieves better detection results than RT-DETR. When dealing with the same mitotic cells, it can display higher confidence levels. Our method successfully identifies subtle lesion patterns that are often missed or misclassified by general-purpose object detection models. Combined with the comparative experiments, this highlights the superior classification accuracy of our model.
Confidence = Pr ( Object ) × IOU truth pred
where Pr ( Object ) reflects the classification accuracy for the “mitotic figure” class, and IOU truth pred represents the localization accuracy between the predicted box and the ground truth.

4. Conclusions

This study provides a systematic analysis of the key challenges in mitosis detection and the limitations of existing algorithms, and innovatively proposes an intelligent detection algorithm based on an Inter–Intra dual-stream architecture. At the feature extraction level, a parallel convolution module is designed to efficiently encode block-level features, and a multi-path gradient propagation mechanism is employed to enhance the representation ability of deep information. In the feature fusion stage, a Bottom–Up reverse refinement strategy is introduced to effectively retain spatial detail information from the original image, constructing a multi-scale feature parsing framework at the inter-level. To address the inadequate modeling of complex intercellular associations, this study introduces, for the first time, the hypergraph neural network (HGNN) into the cell detection domain. Through a learnable hyperedge construction mechanism, HGNN captures the heterogeneous relationships between individual cells and cell groups, forming a higher-order semantic reasoning network at the intra-level. As the first study to apply hypergraph learning to mitosis detection, the experiments show that the proposed method achieves excellent detection performance on public datasets, providing a new technological path for high-throughput mitosis detection.
The limitations of the proposed method and potential remedies are as follows. Imaging and staining conditions vary between different institutions, and a model trained in one dataset may not perform well on another dataset. To address this issue, future research will explore the adoption of an unsupervised staining normalization approach to integrate into the model. In addition, we will investigate the applicability of the proposed detection model to the analysis of pathological images for the diagnosis of other diseases.

Author Contributions

J.L.: Conceptualization and writing—review and editing; Y.Z.: methodology; D.D.: data collation and annotation; G.Z.: data collation and annotation; H.Z.: data collation and annotation; X.X.: paper revision; L.Y.: paper revision and funding acquisition. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Nantong Basic Science Research Program (JC2023021), the Doctoral Research Startup Fund of Nantong University (25B03), and partially by the Qing Lan Project of Jiangsu Province.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data will be made available on request.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Siegel, R.; Miller, K.; Jemal, A. Cancer statistics, 2018. CA Cancer J. Clin. 2018, 68, 7–30. [Google Scholar] [CrossRef]
  2. Wang, N.; Liu, S.; Yang, L.; Zhang, X.; Yuan, Y.; Li, H.; Fu, J.J. Interpretation of 2018 global cancer statistical report. J. Esophageal Surg. (Electron. Vers.) 2019, 5, 87–97. [Google Scholar]
  3. Cree, I.A.; Tan, P.H.; Travis, W.D.; Wesseling, P.; Yagi, Y.; White, V.A.; Lokuhetty, D.; Scolyer, R.A. Counting mitoses: Si (ze) matters! Mod. Pathol. 2021, 34, 1651–1657. [Google Scholar] [CrossRef]
  4. Avallone, G.; Rasotto, R.; Chambers, J.K.; Miller, A.D.; Behling-Kelly, E.; Monti, P.; Berlato, D.; Valenti, P.; Roccabianca, P. Review of histological grading systems in veterinary medicine. Vet. Pathol. 2021, 58, 809–828. [Google Scholar] [CrossRef] [PubMed]
  5. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  6. Jiang, H.; Zhou, Y.; Lin, Y.; Chan, R.C.; Liu, J.; Chen, H. Deep learning for computational cytology: A survey. Med. Image Anal. 2023, 84, 102691. [Google Scholar] [CrossRef] [PubMed]
  7. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef]
  8. Lin, T.-Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal Loss for Dense Object Detection. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2999–3007. [Google Scholar] [CrossRef]
  9. Cireşan, D.; Giusti, A.; Gambardella, L.M.; Schmidhuber, J. Mitosis detection in breast cancer histology images with deep neural networks. In Medical Image Computing and Computer-Assisted Intervention—MICCAI 2013, Proceedings of the 16th International Conference, Nagoya, Japan, 22–26 September 2013; Springer: Berlin/Heidelberg, Germany, 2013; Volume 16, pp. 411–418. [Google Scholar]
  10. Chen, H.; Dou, Q.; Wang, X.; Qin, J.; Heng, P.-A. Mitosis detection in breast cancer histology images via deep cascaded networks. In Proceedings of the Thirtieth AAAI Conference on Artificial Intelligence, AAAI’16, Phoenix, AZ, USA, 12–17 February 2016; AAAI Press: Washington, DC, USA, 2016; pp. 1160–1166. [Google Scholar]
  11. Alom, M.Z.; Aspiras, T.; Taha, T.; Bowen, T.; Asari, V. Mitosisnet: End-to-end mitotic cell detection by multi-task learning. IEEE Access 2020, 8, 68695–68710. [Google Scholar] [CrossRef]
  12. Jahanifar, M.; Shepard, A.; Zamanitajeddin, N.; Bashir, R.S.; Bilal, M.; Khurram, S.A.; Minhas, F.; Rajpoot, N. Stain-robust mitotic figure detection for the mitosis domain generalization challenge. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Strasbourg, France, 27 September–1 October 2021; Springer International Publishing: Cham, Switzerland, 2021; pp. 48–52. [Google Scholar]
  13. Yang, S.; Luo, F.; Zhang, J.; Wang, X. Sk-Unet model with Fourier domain for mitosis detection. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Strasbourg, France, 27 September–1 October 2021; Springer International Publishing: Cham, Switzerland, 2021; pp. 86–90. [Google Scholar]
  14. Scarselli, F.; Gori, M.; Tsoi, A.C.; Hagenbuchner, M.; Monfardini, G. The graph neural network model. IEEE Trans. Neural Netw. 2008, 20, 61–80. [Google Scholar] [CrossRef]
  15. Sun, X.; Xu, C.; Rocha, J.F.; Liu, C.; Hollander-Bodie, B.; Goldman, L.; Krishnaswamy, S.; DiStasio, M.; Perlmutter, M. Hyperedge representations with hypergraph wavelets: Applications to spatial transcriptomics. In Proceedings of the ICASSP 2025—2025 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Hyderabad, India, 6–11 April 2025; IEEE: Piscataway, NJ, USA, 2025; pp. 1–5. [Google Scholar]
  16. Cai, S.; Huang, W.; Yi, W.; Zhang, B.; Liao, Y.; Wang, Q.; Cai, H.; Chen, L.; Su, W. Survival Analysis of Histopathological Image Based on a Pretrained Hypergraph Model of Spatial Transcriptomics Data. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Marrakesh, Morocco, 7–11 October 2024; Springer Nature: Cham, Switzerland, 2024; pp. 455–466. [Google Scholar]
  17. Bretto, A. Hypergraph theory. An introduction. In Mathematical Engineering; Springer: Cham, Switzerland, 2013; Volume 1, pp. 209–216. [Google Scholar]
  18. Feng, Y.; You, H.; Zhang, Z.; Ji, R.; Gao, Y. Hypergraph neural networks. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 1 February 2019; Volume 33, pp. 3558–3565. [Google Scholar]
  19. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  20. Redmon, J.; Farhadi, A. Yolov3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  21. Bochkovskiy, A.; Wang, C.Y.; Liao, H.Y.M. Yolov4: Optimal speed and accuracy of object detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
  22. Jocher, G. Ultralytics YOLOv5. 2020. Available online: https://github.com/ultralytics/yolov5 (accessed on 9 July 2025).
  23. Li, C.; Li, L.; Geng, Y.; Jiang, H.; Cheng, M.; Zhang, B.; Ke, Z.; Xu, X.; Chu, A.X. YOLOv6 v3.0: A full-scale reloading. arXiv 2023, arXiv:2301.05586. [Google Scholar] [CrossRef]
  24. Wang, C.-Y.; Bochkovskiy, A.; Liao, H.-Y.M. YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 7464–7475. [Google Scholar]
  25. Jocher, G.; Chaurasia, A.; Qiu, J. Ultralytics YOLOv8. 2023. Available online: https://github.com/ultralytics/ultralytics (accessed on 9 July 2025).
  26. Wang, C.Y.; Yeh, I.H.; Mark Liao, H.Y. Yolov9: Learning what you want to learn using programmable gradient information. In European Conference on Computer Vision; Springer Nature: Cham, Switzerland, 2024; pp. 1–21. [Google Scholar]
  27. Wang, C.; He, W.; Nie, Y.; Guo, J.; Liu, C.; Wang, Y.; Han, K. Gold-YOLO: Efficient object detector via gather-and-distribute mechanism. In Proceedings of the Thirty-Seventh Conference on Neural Information Processing Systems, New Orleans, LA, USA, 10–16 December 2023. [Google Scholar]
  28. Xu, S.; Wang, X.; Lv, W.; Chang, Q.; Cui, C.; Deng, K.; Wang, G.; Dang, Q.; Wei, S.; Du, Y.; et al. PP-YOLOE: An evolved version of yolo. arXiv 2022, arXiv:2203.16250. [Google Scholar] [CrossRef]
  29. Nazir, A.; Wani, M.A. You only look once-object detection models: A review. In Proceedings of the 2023 10th International Conference on Computing for Sustainable Global Development (INDIACom), New Delhi, India, 15–17 March 2023; IEEE: Piscataway, NJ, USA, 2023; pp. 1088–1095. [Google Scholar]
  30. Li, J.; Zheng, M.; Dong, D.; Xie, X. PC-CS-YOLO: High-Precision Obstacle Detection for Visually Impaired Safety. Sensors 2025, 25, 534. [Google Scholar] [CrossRef]
  31. Tian, Y.; Ye, Q.; Doermann, D. Yolov12: Attention-centric real-time object detectors. arXiv 2025, arXiv:2502.12524. [Google Scholar]
  32. Khanam, R.; Hussain, M. Yolov11: An overview of the key architectural enhancements. arXiv 2024, arXiv:2410.17725. [Google Scholar]
  33. Xu, S.; Fei, J.; Zhao, G.; Liu, X.; Li, H. CCL-YOLO: Catenary Components Location Based on YOLO and Gather-Distribute Mechanism. IEEE Access 2024, 13, 9064–9072. [Google Scholar] [CrossRef]
  34. Dai, J.; Qi, H.; Xiong, Y.; Li, Y.; Zhang, G.; Hu, H.; Wei, Y. Deformable convolutional networks. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 764–773. [Google Scholar]
  35. Wang, K.; Liew, J.H.; Zou, Y.; Feng, J. Panet: Few-shot image semantic segmentation with prototype alignment. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 9197–9206. [Google Scholar]
  36. Aubreville, M.; Stathonikos, N.; Bertram, C.A.; Klopfleisch, R.; ter Hoeve, N.; Ciompi, F.; Wilm, F.; Marzahl, C.; Donovan, T.A.; Maier, A.; et al. Mitosis domain generalization in histopathology images-the MIDOG challenge. Med. Image Anal. 2023, 84, 102699. [Google Scholar] [CrossRef] [PubMed]
  37. Kingma, D.P.; Ba, J. Adam: A Method for Stochastic Optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar] [CrossRef]
  38. Chen, K.; Wang, J.; Pang, J.; Cao, Y.; Xiong, Y.; Li, X.; Sun, S.; Feng, W.; Liu, Z.; Xu, J.; et al. MMDetection: Open MMLab Detection Toolbox and Benchmark. arXiv 2019, arXiv:1906.07155. [Google Scholar] [CrossRef]
  39. Lin, T.Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft COCO: Common Objects in Context. In Lecture Notes in Computer Science, Proceedings of the Computer Vision—ECCV 2014, Zurich, Switzerland, 6–12 September 2014; Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T., Eds.; Springer: Cham, Switzerland, 2014; Volume 8693. [Google Scholar] [CrossRef]
  40. Raschka, S. An overview of general performance metrics of binary classifier systems. arXiv 2014, arXiv:1410.5330. [Google Scholar]
  41. Davis, J.; Goadrich, M. The relationship between Precision-Recall and ROC curves. In Proceedings of the 23rd International Conference on Machine Learning, Pittsburgh, PA, USA, 25–29 June 2006; pp. 233–240. [Google Scholar]
  42. Sun, P.; Zhang, R.; Jiang, Y.; Kong, T.; Xu, C.; Zhan, W.; Tomizuka, M.; Li, L.; Yuan, Z.; Wang, C.; et al. Sparse r-cnn: End-to-end object detection with learnable proposals. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 14454–14463. [Google Scholar]
  43. Cai, Z.; Vasconcelos, N. Cascade r-cnn: Delving into high quality object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 6154–6162. [Google Scholar]
  44. Ge, Z. Yolox: Exceeding yolo series in 2021. arXiv 2021, arXiv:2107.08430. [Google Scholar]
  45. Zhao, Y.; Lv, W.; Xu, S.; Wei, J.; Wang, G.; Dang, Q.; Liu, Y.; Chen, J. Detrs beat yolos on real-time object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16–22 June 2024; pp. 16965–16974. [Google Scholar]
  46. Litjens, G.; Bandi, P.; Ehteshami Bejnordi, B.; Geessink, O.; Balkenhol, M.; Bult, P.; Halilovic, A.; Hermsen, M.; Van de Loo, R.; Vogels, R.; et al. 1399 H&E-stained sentinel lymph node sections of breast cancer patients: The CAMELYON dataset. GigaScience 2018, 7, giy065. [Google Scholar]
  47. Yan, K.; Wang, X.; Lu, L.; Summers, R.M. Deeplesion: Automated deep mining, categorization and detection of significant radiology image findings using large-scale clinical lesion annotations. arXiv 2017, arXiv:1710.01766. [Google Scholar]
Figure 1. Data collection and training workflow. (A) After digitizing the hematoxylin and eosin (H&E)-stained specimens, pathologists annotate the types of mitosis in different tumors. (B) Using a sliding window, the WSI is cropped into patch-level images, which are then fed into a neural network for training to obtain the final results.
Figure 1. Data collection and training workflow. (A) After digitizing the hematoxylin and eosin (H&E)-stained specimens, pathologists annotate the types of mitosis in different tumors. (B) Using a sliding window, the WSI is cropped into patch-level images, which are then fed into a neural network for training to obtain the final results.
Sensors 25 04359 g001
Figure 2. Architecture of the proposed mitosis detection network. The model enhances YOLO11 by introducing Block-Based Mixed Mechanism (BBMM) modules in the backbone, hypergraph neural network (HGNN) in the neck, and a Bottom–Up structure in the head. The network captures both individual cell features and intracell relationships to improve mitosis detection performance.
Figure 2. Architecture of the proposed mitosis detection network. The model enhances YOLO11 by introducing Block-Based Mixed Mechanism (BBMM) modules in the backbone, hypergraph neural network (HGNN) in the neck, and a Bottom–Up structure in the head. The network captures both individual cell features and intracell relationships to improve mitosis detection performance.
Sensors 25 04359 g002
Figure 3. Architecture of the YOLO network modules. The upper part shows the structure of the standard C3K2 module; the lower part shows the Bottleneck module in the proposed Block-Based Mixed Mechanism (BBMM). The redesigned module enhances feature representation and multi-scale fusion capabilities.
Figure 3. Architecture of the YOLO network modules. The upper part shows the structure of the standard C3K2 module; the lower part shows the Bottleneck module in the proposed Block-Based Mixed Mechanism (BBMM). The redesigned module enhances feature representation and multi-scale fusion capabilities.
Sensors 25 04359 g003
Figure 4. Architecture of the BBMM. The BBMM integrates the original C3K2 module with additional 1 × 1 convolution and deformable convolution branches to enhance both inter-channel feature interactions and adaptive spatial modeling.
Figure 4. Architecture of the BBMM. The BBMM integrates the original C3K2 module with additional 1 × 1 convolution and deformable convolution branches to enhance both inter-channel feature interactions and adaptive spatial modeling.
Sensors 25 04359 g004
Figure 5. Architecture of the hypergraph neural network. Feature distances are computed between feature points to construct an incidence matrix and hyperedges, enabling the formation of a hypergraph that models intercellular relationships.
Figure 5. Architecture of the hypergraph neural network. Feature distances are computed between feature points to construct an incidence matrix and hyperedges, enabling the formation of a hypergraph that models intercellular relationships.
Sensors 25 04359 g005
Figure 6. Confusion matrix: (a) baseline; (b) ours.
Figure 6. Confusion matrix: (a) baseline; (b) ours.
Sensors 25 04359 g006
Figure 7. Ablation study on the distance threshold.
Figure 7. Ablation study on the distance threshold.
Sensors 25 04359 g007
Figure 8. Comparative experiment visualization map.
Figure 8. Comparative experiment visualization map.
Sensors 25 04359 g008
Table 1. The results of mitosis detection for different models.
Table 1. The results of mitosis detection for different models.
ModelPrecision (%)Recall (%)mAP50 (%)ACC (%)F1-ScoreParamets (M)GLOPs (G)
Faster R-CNN [7]74.578.282.074.60.76341.3190.1
RetinaNet [8]68.270.572.368.20.69337.9193.8
Sparse R-CNN [42]71.477.180.971.40.741107.3150.7
Cascade R-CNN [43]79.385.188.479.30.82177.3278.4
YOLOX [44]73.778.286.173.70.75947.1115.6
YOLO11 [32]77.481.687.677.40.79425.386.9
RT-DETR [45]80.384.587.280.30.82342136
Ours83.189.793.683.60.86356.3211
Table 2. Result of ablation experiments.
Table 2. Result of ablation experiments.
ModelBBMMHGNNPrecision (%)Recall (%)mAP50 (%)
Baseline [32] 77.481.687.5
Baseline + BBMM 80.486.988.2
Baseline + HGNN 79.187.390.1
Ours83.189.793.6
Note: ‘✓’ indicates that a corresponding improvement has been made.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, J.; Dong, D.; Zhan, Y.; Zhu, G.; Zhang, H.; Xie, X.; Yang, L. Cross-Scale Hypergraph Neural Networks with Inter–Intra Constraints for Mitosis Detection. Sensors 2025, 25, 4359. https://doi.org/10.3390/s25144359

AMA Style

Li J, Dong D, Zhan Y, Zhu G, Zhang H, Xie X, Yang L. Cross-Scale Hypergraph Neural Networks with Inter–Intra Constraints for Mitosis Detection. Sensors. 2025; 25(14):4359. https://doi.org/10.3390/s25144359

Chicago/Turabian Style

Li, Jincheng, Danyang Dong, Yihui Zhan, Guanren Zhu, Hengshuo Zhang, Xing Xie, and Lingling Yang. 2025. "Cross-Scale Hypergraph Neural Networks with Inter–Intra Constraints for Mitosis Detection" Sensors 25, no. 14: 4359. https://doi.org/10.3390/s25144359

APA Style

Li, J., Dong, D., Zhan, Y., Zhu, G., Zhang, H., Xie, X., & Yang, L. (2025). Cross-Scale Hypergraph Neural Networks with Inter–Intra Constraints for Mitosis Detection. Sensors, 25(14), 4359. https://doi.org/10.3390/s25144359

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop