Next Article in Journal
AI Based Clinical Decision-Making Tool for Neurologists in the Emergency Department
Previous Article in Journal
Antiplatelet Therapy in Atrial Fibrillation Patients on Direct Oral Anticoagulants Undergoing Percutaneous Coronary Intervention: Which and How
Previous Article in Special Issue
Real-Life Performance of a Commercially Available AI Tool for Post-Traumatic Intracranial Hemorrhage Detection on CT Scans: A Supportive Tool
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Artificial Intelligence Algorithm Supporting the Diagnosis of Developmental Dysplasia of the Hip: Automated Ultrasound Image Segmentation

1
Department of Orthopedics and Traumatology, Medical University of Warsaw, Lindley 4 Str., 02-005 Warsaw, Poland
2
Pentacomp Systemy Informatyczne S.A., Aleje Jerozolimskie 179 Str., 02-222 Warsaw, Poland
3
Gustav—Children’s Clinic, Białej Floty 2 Str., 02-654 Warsaw, Poland
*
Author to whom correspondence should be addressed.
J. Clin. Med. 2025, 14(17), 6332; https://doi.org/10.3390/jcm14176332
Submission received: 31 July 2025 / Revised: 27 August 2025 / Accepted: 5 September 2025 / Published: 8 September 2025

Abstract

Background: Developmental dysplasia of the hip (DDH), if not treated, can lead to osteoarthritis and disability. Ultrasound (US) is a primary screening method for the detection of DDH, but its interpretation remains highly operator-dependent. We propose a supervised machine learning (ML) image segmentation model for the automated recognition of anatomical structures in hip US images. Methods: We conducted a retrospective observational analysis based on a dataset of 10,767 hip US images from 311 patients. All images were annotated for eight key structures according to the Graf method and split into training (75.0%), validation (9.5%), and test (15.5%) sets. Model performance was assessed using the Intersection over Union (IoU) and Dice Similarity Coefficient (DSC). Results: The best-performing model was based on the SegNeXt architecture with an MSCAN_L backbone. The model achieved high segmentation accuracy (IoU; DSC) for chondro-osseous border (0.632; 0.774), femoral head (0.916; 0.956), labrum (0.625; 0.769), cartilaginous (0.672; 0.804), and bony roof (0.725; 0.841). The average Euclidean distance for point-based landmarks (bony rim and lower limb) was 4.8 and 4.5 pixels, respectively, and the baseline deflection angle was 1.7 degrees. Conclusions: This ML-based approach demonstrates promising accuracy and may enhance the reliability and accessibility of US-based DDH screening. Future applications could integrate real-time angle measurement and automated classification to support clinical decision-making.

1. Introduction

Developmental dysplasia of the hip (DDH) is the most common musculoskeletal disorder in newborns [1]. It affects approximately 1.4% (95% CI: 0.86 to 2.28) of the world’s population, but the incidence of DDH can vary from 0.5% to 30% depending on geographical and ethnic origin [2]. The term DDH reflects a broad spectrum of abnormalities, from mild acetabular deficiency to severe cases with dislocation of the femoral head. In less severe cases, DDH can remain clinically silent in children due to the limited sensitivity and specificity of clinical examination findings, such as hip abduction difference, asymmetric skin creases, or hip instability tests. If not treated, it often leads to the development of hip osteoarthritis (OA) in young adults, which may require total hip arthroplasty (THA) [3]. Up to 40% of patients who undergo THA under 50 years of age have radiological signs of DDH [4]. In severe DDH, symptoms usually appear at birth, including hip instability, shortening of the limb, and limited abduction of the affected hip. In children, we can observe gait problems or a delay in the onset of walking [5]. Conservative treatment (bracing) has the highest success rate if DDH is diagnosed during the first 6 weeks of life. When the diagnosis is delayed or incorrect, surgery may be necessary [6].
Before the introduction of hip ultrasound (US) by Reinhard Graf (1980s), the diagnosis of DDH was based on clinical examination and radiographic imaging. As mentioned above, clinical examination has only limited efficacy in the detection of DDH, and radiograph-based diagnosis is only possible after 4 to 6 months of age, which is too late for efficient conservative treatment [7]. Today, ultrasound is the gold standard in the diagnosis and monitoring of DDH treatment [8]. Graf’s method evaluates acetabulum development by measuring the α angle on standard plane US hip scans. Depending on the α angle, hips are classified (I-IV), and the β angle determines the hip subtype. Graf’s method is characterized by high sensitivity and specificity in DDH detection, respectively, at 93% (95% CI 0.57–0.99) and 97% (95% CI 0.86–0.99) [9]. However, many errors in the application of the method are reported in the literature, and the use of the Graf method requires experience and strict adherence to the usability checklists [10]. In Europe, many countries have adopted universal ultrasound screening, resulting in a high demand for specialists in the field of DDH diagnosis [8].
The solution for the large volume of US hip scans required could be the recently developed ML-based algorithms. They can improve the reliability of the diagnosis, which is often compromised by the relatively poor quality of the images due to the presence of noise and acquisition errors. They can provide a “second opinion” in diagnosis by standardizing image interpretation and reducing operator-dependent errors [11]. Various diagnostic applications of AI in musculoskeletal ultrasound already include the detection, diagnosis, and prognosis of diseases [12]. The precision of such an application of AI in hip US could equal or even outperform that of humans. Some AI-based systems to detect DDH have already been approved by the FDA [13]. However, none of the systems are fully based on the widely accepted Graf rules of image acquisition. In our study, we present a system for automated identification of anatomical structures and landmarks proposed by Graf in two-dimensional (2D) US images. Subsequently, this system could be used in clinical applications to calculate in real time the α and β angles and suggest the possible diagnosis of DDH.
The purpose of this study is to address the growing need for reliable, high-volume ultrasound screening of infant hips by introducing a machine learning-based system that standardizes image interpretation. Current AI tools for DDH detection are limited, as none are fully aligned with the widely accepted Graf methodology, which remains the clinical gold standard. By developing an automated solution capable of identifying anatomical structures and landmarks required by Graf, this approach has the potential to reduce operator dependency, improve diagnostic consistency, and facilitate large-scale, efficient screening for DDH in future applications.

2. Materials and Methods

2.1. Study Materials

This was a retrospective cross-sectional observational study. The material for image annotation was collected at a private orthopedic clinic during universal screening US scans (4–6 weeks of life) and follow-up visits. All examinations were performed according to the Graf method by six different orthopedic surgeons trained in this method. The data were acquired in the period between January 2022 and December 2023.
The sample material comes from Mindray USG DC-60S (Shenzhen Mindray Bio-Medical Electronics Co., Ltd., Shenzhen, China), and the settings are individually adjusted by each examiner. Videos (DICOM files) were stored on the PACS server (Mini PACS server HP ProLiant DL20 G10 (Hewlett Packard Enterprise, Houston, TX, USA) with processor: E-2224, RAM: 16 GB) equipped with Mini mEdivum Electronic Ultrasound Examination Archiving System (mEdivum Sp. z o.o., Warsaw, Poland). All data on DICOM files were anonymized using dedicated automated tool.
The structures were annotated by three independent orthopedic surgeons trained in the Graf method who had participated in a Graf-accredited course. The contours of the anatomical structures in the US images were marked using points along the edges of the target anatomy by dedicated software on selected frames. In the annotation process, a maximum of 25 frames from each video were selected by the reporting physician as the best. Annotating physicians selected structures for marking from a drop-down list among the objects on Checklists I and II according to Prof. Graf, necessary for the assessment of hip joints for DDH (Figure 1). Frames were selected for annotation only when it was possible to mark 5 or more objects. Furthermore, at least one selected frame in each video was classified as diagnostic, according to the Graf methodology (standard plane) [14].
To minimize potential sources of bias, several measures were undertaken. One of the primary ways to reduce the risk of error involves real-time automated feedback. The annotation tool identifies common errors, such as overlap or incorrect region selection in ultrasound images. To reduce subjectivity, each annotated frame was independently reviewed by two physicians (cross-check). The professionals were trained not only to use the tool, but also to learn from mistakes made in previous batches of annotations. This involved learning from the feedback of independent reviewers and documenting common errors in the form of a shared knowledge base. The study size was determined pragmatically by including all available scans; this dataset is larger than in most previous AI-DDH studies.

2.2. Artificial Intelligence Model

In our approach, the AI model is trained using supervised learning, which usually allows for the best result, but is very demanding regarding the data gathering and labeling process. The model is presented with a US image from a hip scan and returns a segmentation mask with key anatomical structures that contains five structures and two critical points, marked as squares with a size of 5 pixels and a baseline marked as a segment of 3-pixel thickness. Deep neural networks with gradient descent optimization were used. The specifications of the model architectures are described in a later section.
Two approaches to segmentation were tested. In the first approach (Model-5), data for 5 structures were fed into a segmentation model based on convolutional layers and attention mechanisms. In the second approach (Model-8), the data for five structures were augmented with two points (lower limb and bony rim) and the baseline. In both models, five different architectures were tested, and the best one was selected. We tested the following: SegFormer [15], OCRNet [2], U-Net [16], U-HRNet [17], and SegNeXt [18] (Supplementary Table S1). For each architecture, there were many training settings (described in Section 3). Among the many architectures and settings, the one with the best Intersection over Union (IoU) was used.
Additionally, during the model training process, the PaddleSeg library uses various pre-processing techniques (e.g., Resize, Normalize) and augmentation techniques (e.g., RandomHorizontalFlip, RandomDistort), which were also used in our training. The first step of pre-processing involved cropping the images to the ultrasound area to eliminate regions that do not pertain directly to the ultrasound scan. The second step involved the preparation of masks containing only 5 classes (chondro-osseous border, femoral head, labrum, cartilaginous roof, and bony roof). This step was performed to create the input for Model-5 (Figure 2).
For segmentation labels from models, post-processing was used, which removed artifacts related to multiple instances of a single class caused by either small regions that needed to be cleaned or divided areas that should be connected. Moreover, mistakes for regions that were not properly marked by the model were also corrected. The best image had to be selected for post-processing. Selection was made based on a scoring algorithm that evaluated the quality of the segmentation mask and the correctness of the US image itself, whether it had all the necessary classes visible, and whether it presented a Graf standard plane.

2.3. Statistical Methods

The evaluation of models is crucial for understanding their effectiveness. In the field of image segmentation, one of the most widely adopted metrics is Intersection over Union (IoU). It quantifies the overlap between the predicted segmentation mask and the ground truth mask for a specific object class. Mathematically, IoU is defined as the ratio of the area of intersection between the predicted and ground truth regions to the area of their union. Formally, for a given class C, the IoU can be expressed as:
I o U C =   | A B | | A B |
where A represents the area of the predicted segmentation mask and B denotes the area of the ground truth mask. The intersection ∣AB∣ represents the area where both masks overlap, while the union ∣AB∣ represents the total area covered by at least one of the masks. A higher IoU value indicates better model performance. To evaluate the models, the mean Intersection over Union for a specific class across an entire dataset was computed. Another metric we used is the dice similarity coefficient (DSC), which ranges from 0 to 1. Here, A represents the predicted point set and B the labeled point set.
D S C ( S A , S P ) = 2   ·   | A   B | | A | + | B |
All images were split into three sets: training set (for model learning), validation set (to select the best training iteration), and test set (to choose the best model among tested architectures and for final model evaluation). Images related to the same patient had to appear in only one of the sets. The split was performed in such a way that the test and validation sets each contained approximately 10% of the videos (Table 1).
Given the dependency between paired observations and non-normal distribution, the Wilcoxon signed-rank test, a non-parametric alternative to the paired t-test, was employed. To control for the family-wise error rate resulting from multiple comparisons, the Bonferroni correction was applied, with the significance threshold set at α = 0.01. All statistical analyses were performed using Python version 3.9.13 with the SciPy library (version 1.7.3).

3. Results

The dataset consisted of 375 female and 310 male US hip scans. Dysplastic hips (DDH) or physiologically immature hips (type IIa+) were diagnosed in 106 (15.47%) of all 685 DICOM files. The descriptive statistics for the dataset are presented in Table 2.
The accuracy of the model was evaluated using the IoU metric. The average value of this metric was calculated based on the test dataset for five classes (chondro-osseous border, femoral head, labrum, cartilaginous roof, and bony roof). For classes representing points (bony rim and lower limb), the Euclidean distance from the centroid of the model’s segment to the centroid of the doctor’s segment was calculated. For the baseline class, the angle of deviation between the doctor’s baseline and the model’s baseline was computed. Two models were analyzed: Model-8 (generating segments for eight classes) and Model-5 (generating segments for five classes). The analysis revealed statistically significant differences between the models in classes 1, 3, and 4. In class 1, the 8-class model demonstrated superior performance, whereas in classes 3 and 4, the 5-class model achieved higher IoU scores compared to the 8-class model. Despite reaching statistical significance, the observed effect sizes ranged from small to moderate, indicating only moderate practical relevance of these differences. The average values of the metrics calculated for the test dataset for each class are presented in Table 3.
To evaluate the effectiveness of the segmentation model, the success rate was calculated, defined as the percentage of frames in which the doctor annotated a given class and for which the Intersection over Union (IoU) between the doctor’s annotation and the model prediction reaches a value of at least 0.5 (true positives). The results are presented in Table 4.
Furthermore, for five classes (chondro-osseous border, femoral head, labrum, cartilaginous roof, and bony roof), the dice similarity coefficient on the test dataset was calculated. The average values of the metrics for each class are presented in Table 5.
The model that achieved the highest segmentation accuracy was based on the SegNeXt architecture with the MSCAN_L backbone. SegNeXt is a convolutional network that features hierarchical attention mechanisms (HAM), which contribute to its high segmentation accuracy [18]. The training parameters of our best model are detailed in Supplementary Table S2.
To illustrate the performance of the segmentation model, a comparison is presented between the labels obtained by the segmentation model and the ground truth labels. Two models were analyzed: Model-8 (generating segments for eight classes) and Model-5 (generating segments for five classes). Both the cases with the highest quality model labels and the cases where the model labels were less accurate are presented. Figure 3 shows one of the best cases, in which Model-5 and Model-8 accurately labeled the classes relative to the doctor’s labels. The worst-case scenario was also analyzed, in which the prediction results differ significantly from the ground truth labels (Figure 4).

4. Discussion

In our work, orthopedic surgeons annotated all the structures recommended by Graf to obtain reproducible results. We expect that the structures obtained in this way can be used to make a reliable clinical diagnosis regarding the type of hip joint in the future. The SegNeXt MSCAN_L model achieved high accuracy, especially for the femoral head (IoU 0.916; DSC 0.956) and bony roof (IoU 0.725; DSC 0.841), with reliable performance for other key structures and landmarks. These results demonstrate that the system can replicate expert annotations, reduce operator dependency, and support standardized, automated DDH screening.
When compared with previous research, our approach differs in both methodology and scope, yet the achieved accuracy is consistent with or superior to reported values. Some studies, such as Quander et al. and Hareendranathan et al., use a different approach and analyze the geometry of structures instead of supervised learning methods [19]. Multiple authors analyze 3D US scans, but the methodology proposed by Graf is based on 2D US [20,21]. According to the literature, the sensitivity [95% CI] of such AI US-based systems can be as high as 0.98 [0.91–1.00]. Their specificity [95% CI] is also reported to be high—0.95 [0.81–0.99] [22].
Sezer et al. used 675 annotated US frames with “iliac wing”, “labrum”, and acetabulum. The R-CNN (Mask Region-Based Convolutional Neural Network) method was used. The authors do not report the results as an IoU metric. Instead, they report a success rate that is IoU of ≥0.5 (true positive). The average success rate was 98.25% for the iliac wing, 94.91% for the acetabulum, and 97.72% for the labrum [23]. Our success rate for the bony roof, which represents the acetabulum and iliac wing, ranged from 97.85% (Model-8) to 98.03% (Model-5). For the labrum, our model had a success rate ranging from 82.41% (Model-8) to 85.76% (Model-5).
Chen et al. described the use of an improved fully convolutional neural network to identify the femoral head on US images. The raw dataset consisted of 55 labeled US images of hip joints. The authors used the Cascaded FNet method and extracted a region of interest (ROI). The model achieved an IoU of 0.897 [24]. Our models achieved an IoU for the femoral head ranging from 0.915 (Model-5) to 0.916 (Model-8).
Li et al. used 400 US Graf standard plane images for a semi-supervised deep learning method based on a feature pyramid network (FPN) and a contrastive learning scheme based on a Siamese architecture. The annotations included the junction of cartilage and bone (chondro-osseous border), hyaline cartilage (cartilaginous roof), acetabular roof (bony roof), femoral head, joint capsule, labrum, ilium, bony part of the acetabular roof, and synovial fold. However, the authors did not include the lower limb and bony rim, which are necessary for US scan assessment according to Graf [25]. For the chondro-osseous border, the authors achieved DSC ranging from 0.685 to 0.864. Our model DSC ranged from 0.769 (Model-5) to 0.774 (Model-8). For the femoral head, the authors achieved a DSC of 0.860–0.899, and our result was 0.956 (both Model-5 and Model-8). For the labrum, the DSC in Li et al. was 0.570–0.814, and our result was 0.767 (Model-8)–0.769 (Model-5). For the cartilaginous roof, the DSC in Li et al. was 0.000–0.565, while our results were 0.799 (Model-8)–0.804 (Model-5). For the ilium and lower ilium, the authors’ results were 0.6802–0.854 and 0.379–0.819, respectively, and our result for the bony roof that combines both structures was 0.834 (Model-8)–0.841 (Model-5).
Stamper et al. was based on 190 US hip scans. The annotations included the femoral head, ilium, and labrum. U-Net was used to segment key anatomical structures. However, this study was aimed at automating femoral head coverage (FHC) calculation for DDH screening, and ours was focused on the Graf method [26]. The authors reported DSC for the femoral head of 0.924, and our result was 0.956 (both Model-8 and Model-5). The result for the ilium was 0.857, and ours for the bony roof was 0.834 (Model-8)–0.841 (Model-5). For the labrum, Stamper et al. achieved a DSC of 0.710; our result was 0.767 (Model-8) and 0.769 (Model-5).
Hu et al. used Mask R-CNN on 1231 US images of the infant hip from 632 patients and annotated the flat ilium, lower limb, labrum, CO junction (chondro-osseous border), bony rim, lower limb point, and midpoint of labrum. For the chondro-osseous border, the authors achieved a DSC ranging from 0.829 to 0.873 [27]. Our model DSC ranged from 0.769 (Model-5) to 0.774 (Model-8). For the labrum, the DSC of Hu et al. was 0.791–0.841, and our result was 0.767 (Model-8)–0.769 (Model-5). For the ilium and lower limb (marked as an area and part of the bony roof), the results were 0.869–0.869 and 0.809–0.838, respectively, and our result for the bony roof that combines both structures was 0.834 (Model-8)–0.841 (Model-5).
Golan et al. used the DCNN network and GAN. The authors chose a unique approach in which 1056 US scans of infant hips were annotated by crowdsourcing to crowdFlower platform users. The annotated structures were the ilium and acetabular roof, but the authors do not report DSC or the IoU metric [28].
Lee et al. analyzed 1243 hip US images from 168 infants. Mask annotations were made using the Computer Vision Annotation Tool (CVAT). The annotated structure was a combined area of bony roof and cartilaginous roof marked on a Graf standard plane. The authors do not report the DSC or IoU [29].
A limitation of this study is the fixed nature of the data splits and the inability to perform repeated training-validation cycles. This was a necessary methodological consequence of the project’s design, where data were acquired sequentially. Early data tranches were used for the iterative development and refinement of the pre- and post-processing algorithms. To ensure an unbiased assessment of the final model, only the last-received, entirely unseen dataset could be used for testing, precluding resampling for multiple runs or for balancing dataset ratios. Furthermore, the observed gender imbalance within the test set reflects the demographic characteristics of this final data batch and was a factor beyond our control.
The generalizability of our findings may be limited by the single-center nature of this retrospective study, as all data were obtained from one orthopedic clinic using a single ultrasound system. Differences in ultrasound equipment, examiner experience, and patient demographics across other institutions could influence model performance. At the same time, the inclusion of both universal screening and follow-up examinations reflects routine clinical practice and enhances the potential applicability of the results. To confirm robustness and broader external validity, future multicenter studies involving diverse imaging systems, operators, and populations will be required.
In our study, we developed an AI system that, for the first time, can automatically recognize all the anatomical structures required for the Graf method of hip ultrasound assessment. In practice, this means that the algorithm can “see” the same landmarks that a clinician identifies manually in order to measure the α and β angles. As a result, the calculation of these key diagnostic parameters can be performed automatically and consistently. This could reduce operator-dependent variability, help in interpreting images of lower quality, and facilitate large-scale screening programs. Importantly, the accuracy of our method, measured by IoU, DSC, and success rate, is comparable to or even better than that reported in previous studies on AI in hip ultrasound.
In future studies, we plan to design a complete CDSS class system (Clinical Decision Support System) to support doctors in their complex decision-making processes regarding DDH diagnosis and management. Recently, CDSS have seen a rapid evolution, and they are now commonly administered through computerized clinical workflows and medical records [30]. The system will be based on a machine learning (ML) algorithm that measures the α and β angles in strict accordance with the Graf method. Other data, such as patient demographics, potential risk factors for DDH, and clinical examination results, will also be collected and analyzed. Further benefits of artificial intelligence (AI) in ultrasound imaging may include improved diagnostic accuracy, leading to better patient outcomes.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/jcm14176332/s1, Table S1. Architectures tested in the study; Table S2. Model training configuration.

Author Contributions

Conceptualization: Ł.P. and J.K.; Methodology: P.C., B.M., M.P. and J.W.; Software: P.C., B.M., M.P. and J.W.; Resources: Ł.P., P.Ł. and J.K.; Writing—original draft preparation: Ł.P., P.C., J.K., B.M., M.P. and J.W.; Writing—review and editing: Ł.P., P.C., J.K., B.M., M.P., P.Ł. and J.W.; Visualization: B.M., M.P. and Ł.P.; Supervision: P.C. and Ł.P.; Project administration: J.K. and P.C.; Funding acquisition: J.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Polish Medical Research Agency (ABM), grant number 2022/ABM/02/00004. The Article Publication Charges were funded by the Medical University of Warsaw.

Institutional Review Board Statement

This study was conducted in accordance with the Declaration of Helsinki. The Ethics Committee of the Medical University of Warsaw (protocol code: AKBE/07/2022 and date: 17 January 2022) approved the study protocol and waived the requirement for informed consent due to its retrospective design.

Informed Consent Statement

Not applicable, retrospective observational study.

Data Availability Statement

The data that support the findings of this study are available from Pentacomp Systemy Informatyczne S.A., but restrictions apply to the availability of these data, which were used under license for the current study, and so are not publicly available. Data are, however, available from the authors upon reasonable request and with permission of Pentacomp Systemy Informatyczne S.A.

Conflicts of Interest

The co-authors are employees (Paweł Czech, Jadwiga Kaliszewska, Bartłomiej Mulewicz, Maciej Pykosz, Wiszniewska Joanna) or cooperate (Łukasz Pulik, Jadwiga Kaliszewska) with Pentacomp Systemy Informatyczne S.A., which may benefit from the development of products as an outcome of the work. The authors state no other conflicts of interest.

References

  1. Pulik, Ł.; Płoszka, K.; Romaniuk, K.; Sibilska, A.; Jedynak, A.; Tołwiński, I.; Kumięga, P.; Wojtyński, P.; Łęgosz, P. Impact of Multiple Factors on the Incidence of Developmental Dysplasia of the Hip: Risk Assessment Tool. Medicina 2022, 58, 1158. [Google Scholar] [CrossRef] [PubMed]
  2. Tao, A.; Sapra, K.; Catanzaro, B. Hierarchical multi-scale attention for semantic segmentation. arXiv 2020, arXiv:2005.10821. [Google Scholar] [CrossRef]
  3. Subaşı, İ.Ö.; Veizi, E.; Çepni, Ş.; Alkan, H.; Oğuz, T.; Fırat, A. Clinical Examination Findings Can Accurately Diagnose Developmental Dysplasia of The Hip—A Large, Single-Center Cohort. Children 2023, 10, 304. [Google Scholar] [CrossRef]
  4. Muddaluru, V.; Boughton, O.; Donnelly, T.; O’Byrne, J.; Cashman, J.; Green, C. Developmental dysplasia of the hip is common in patients undergoing total hip arthroplasty under 50 years of age. SICOT-J 2023, 9, 25. [Google Scholar] [CrossRef]
  5. Anable, N.R.; Luginsland, L.A.; Carlos, C.; Stevens, W.R., Jr.; Loewen, A.M.; Jeans, K.A.; Sucato, D.J. Investigating pelvic drop gait abnormality in adolescent hip pathology patients. Gait Posture 2024, 110, 65–70. [Google Scholar] [CrossRef]
  6. He, J.; Chen, T.; Lyu, X. Analysis of the results of hip ultrasonography in 48 666 infants and efficacy studies of conservative treatment. J. Clin. Ultrasound 2023, 51, 656–662. [Google Scholar] [CrossRef]
  7. Kitay, A.; Widmann, R.F.; Doyle, S.M.; Do, H.T.; Green, D.W. Ultrasound is an Alternative to X-ray for Diagnosing Developmental Dysplasia of the Hips in 6-Month-Old Children. HSS J. 2019, 15, 153–158. [Google Scholar] [CrossRef]
  8. Krysta, W.; Dudek, P.; Pulik, Ł.; Łęgosz, P. Screening of Developmental Dysplasia of the Hip in Europe: A Systematic Review. Children 2024, 11, 97. [Google Scholar] [CrossRef]
  9. Chavoshi, M.; Mirshahvalad, S.A.; Mahdizadeh, M.; Zamani, F. Diagnostic Accuracy of Ultrasonography Method of Graf in the detection of Developmental Dysplasia of the Hip: A Meta-Analysis and Systematic Review. Arch. Bone Jt. Surg. 2021, 9, 297–305. [Google Scholar] [CrossRef]
  10. Ossendorff, R.; Placzek, S.; Bornemann, R.; Walter, S.G. Four decades of the Graf method in screening for developmental dysplasia of the hip (part I): Rightly the gold standard or of dubious benefit? Front. Pediatr. 2022, 10, 1002696. [Google Scholar] [CrossRef] [PubMed]
  11. Micucci, M.; Iula, A. Recent Advances in Machine Learning Applied to Ultrasound Imaging. Electronics 2022, 11, 1800. [Google Scholar] [CrossRef]
  12. Hattori, S.; Saggar, R.; Heidinger, E.; Qi, A.; Mullen, J.; Fee, B.; Brown, C.L.; Canton, S.P.; Scott, D.; Hogan, M.V. Advances in Ultrasound-Guided Surgery and Artificial Intelligence Applications in Musculoskeletal Diseases. Diagnostics 2024, 14, 2008. [Google Scholar] [CrossRef] [PubMed]
  13. Jaremko, J.L.; Hareendranathan, A.; Bolouri, S.E.S.; Frey, R.F.; Dulai, S.; Bailey, A.L. AI aided workflow for hip dysplasia screening using ultrasound in primary care clinics. Sci. Rep. 2023, 13, 9224. [Google Scholar] [CrossRef] [PubMed]
  14. Graf, R.; Mohajer, M.; Plattner, F. Hip sonography update. Quality-management, catastrophes—Tips and tricks. Med. Ultrason. 2013, 15, 299–303. [Google Scholar] [CrossRef]
  15. Xie, E.; Wang, W.; Yu, Z.; Anandkumar, A.; Alvarez, J.M.; Luo, P. SegFormer: Simple and efficient design for semantic segmentation with transformers. In Proceedings of the 35th International Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 10–15 December 2024; p. 924. [Google Scholar]
  16. Huang, H.; Lin, L.; Tong, R.; Hu, H.; Zhang, Q.; Iwamoto, Y.; Han, X.; Chen, Y.-W.; Wu, J. Unet 3+: A full-scale connected unet for medical image segmentation. In Proceedings of the ICASSP 2020—2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Virtual, 4–9 May 2020; pp. 1055–1059. [Google Scholar]
  17. Wang, J.; Long, X.; Chen, G.; Wu, Z.; Chen, Z.; Ding, E. U-hrnet: Delving into improving semantic representation of high resolution network for dense prediction. arXiv 2022, arXiv:2210.07140. [Google Scholar] [CrossRef]
  18. Zhang, Y.; Yao, T.; Qiu, Z.; Mei, T. Lightweight and progressively-scalable networks for semantic segmentation. Int. J. Comput. Vis. 2023, 131, 2153–2171. [Google Scholar] [CrossRef]
  19. Quader, N.; Hodgson, A.J.; Mulpuri, K.; Schaeffer, E.; Abugharbieh, R. Automatic Evaluation of Scan Adequacy and Dysplasia Metrics in 2-D Ultrasound Images of the Neonatal Hip. Ultrasound Med. Biol. 2017, 43, 1252–1262. [Google Scholar] [CrossRef]
  20. Quader, N.; Hodgson, A.; Mulpuri, K.; Cooper, A.; Abugharbieh, R. Towards reliable automatic characterization of neonatal hip dysplasia from 3D ultrasound images. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Athens, Greece, 17–21 October 2016; pp. 602–609. [Google Scholar]
  21. Paserin, O.; Mulpuri, K.; Cooper, A.; Hodgson, A.J.; Garbi, R. Real time RNN based 3D ultrasound scan adequacy for developmental dysplasia of the hip. In Proceedings of the Medical Image Computing and Computer Assisted Intervention–MICCAI 2018: 21st International Conference, Granada, Spain, 16–20 September 2018; Springer: Berlin/Heidelberg, Germany, 2018. Proceedings, Part I. [Google Scholar]
  22. Chen, M.; Cai, R.; Zhang, A.; Chi, X.; Qian, J. The diagnostic value of artificial intelligence-assisted imaging for developmental dysplasia of the hip: A systematic review and meta-analysis. J. Orthop. Surg. Res. 2024, 19, 522. [Google Scholar] [CrossRef]
  23. Sezer, A.; Sezer, H.B. Segmentation of measurable images from standard plane of Graf hip ultrasonograms based on Mask Region-Based Convolutional Neural Network. Jt. Dis. Relat. Surg. 2023, 34, 590. [Google Scholar] [CrossRef]
  24. Chen, L.; Cui, Y.; Song, H.; Huang, B.; Yang, J.; Zhao, D.; Xia, B. Femoral head segmentation based on improved fully convolutional neural network for ultrasound images. Signal Image Video Process. 2020, 14, 1043–1051. [Google Scholar] [CrossRef]
  25. Li, X.; Zhang, R.; Wang, Z.; Wang, J. Semi-supervised learning in diagnosis of infant hip dysplasia towards multisource ultrasound images. Quant. Imaging Med. Surg. 2024, 14, 3707. [Google Scholar] [CrossRef]
  26. Stamper, A.; Singh, A.; McCouat, J.; Voiculescu, I. Infant hip screening using multi-class ultrasound scan segmentation. In Proceedings of the 2023 IEEE 20th International Symposium on Biomedical Imaging (ISBI), Cartagena, Colombia, 18–21 April 2023; pp. 1–4. [Google Scholar]
  27. Hu, X.; Wang, L.; Yang, X.; Zhou, X.; Xue, W.; Cao, Y.; Liu, S.; Huang, Y.; Guo, S.; Shang, N. Joint landmark and structure learning for automatic evaluation of developmental dysplasia of the hip. IEEE J. Biomed. Health Inform. 2021, 26, 345–358. [Google Scholar] [CrossRef]
  28. Golan, D.; Donner, Y.; Mansi, C.; Jaremko, J.; Ramachandran, M.; CUDL. Fully automating Graf’s method for DDH diagnosis using deep convolutional neural networks. In Proceedings of the Deep Learning and Data Labeling for Medical Applications: First International Workshop, LABELS 2016, and Second International Workshop, DLMIA 2016, Held in Conjunction with MICCAI 2016, Athens, Greece, 21 October 2016; Proceedings 1. pp. 130–141. [Google Scholar]
  29. Lee, S.-W.; Ye, H.-U.; Lee, K.-J.; Jang, W.-Y.; Lee, J.-H.; Hwang, S.-M.; Heo, Y.-R. Accuracy of new deep learning model-based segmentation and key-point multi-detection method for ultrasonographic developmental dysplasia of the hip (DDH) screening. Diagnostics 2021, 11, 1174. [Google Scholar] [CrossRef]
  30. Sutton, R.T.; Pincock, D.; Baumgart, D.C.; Sadowski, D.C.; Fedorak, R.N.; Kroeker, K.I. An overview of clinical decision support systems: Benefits, risks, and strategies for success. NPJ Digit. Med. 2020, 3, 17. [Google Scholar] [CrossRef]
Figure 1. Ultrasound image of the hip in Graf standard plane (a) and labeling of anatomical structures (b). The labeled diagram (b) includes 5 anatomical structures, i.e., chondro-osseous border (red-1), femoral head (purple-2), cartilaginous roof (yellow-3), bony roof (beige-4), and labrum (light purple-5). Other landmarks include lower limb (orange-6), bony rim (light blue-7), and baseline (pink-8).
Figure 1. Ultrasound image of the hip in Graf standard plane (a) and labeling of anatomical structures (b). The labeled diagram (b) includes 5 anatomical structures, i.e., chondro-osseous border (red-1), femoral head (purple-2), cartilaginous roof (yellow-3), bony roof (beige-4), and labrum (light purple-5). Other landmarks include lower limb (orange-6), bony rim (light blue-7), and baseline (pink-8).
Jcm 14 06332 g001
Figure 2. Result of pre-processing: extracting the USG area from the image and preparing masks for Model-5 (lower picture) and Model-8 (upper picture).
Figure 2. Result of pre-processing: extracting the USG area from the image and preparing masks for Model-5 (lower picture) and Model-8 (upper picture).
Jcm 14 06332 g002
Figure 3. High-quality model labels (upper—Model-8) and (lower—Model-5). Annotated mask on the left, model mask on the right.
Figure 3. High-quality model labels (upper—Model-8) and (lower—Model-5). Annotated mask on the left, model mask on the right.
Jcm 14 06332 g003
Figure 4. Low-quality model labels (upper—Model-8) and (lower—Model-5). Annotated mask on the left, model mask on the right.
Figure 4. Low-quality model labels (upper—Model-8) and (lower—Model-5). Annotated mask on the left, model mask on the right.
Jcm 14 06332 g004
Table 1. Numerical and percentage share of data sets (with column percentages).
Table 1. Numerical and percentage share of data sets (with column percentages).
Videos n (%)Images n (%)
Test set93 (13.6%)1671 (15.5%)
Validation set72 (10.5%)1022 (9.5%)
Training set520 (75.9%)8074 (75.0%)
Table 2. Descriptive statistics of study population (with row percentages).
Table 2. Descriptive statistics of study population (with row percentages).
Age, Days *Female (%)
Test set49 (40–75)70 (75.3%)
Validation set58 (40–79)40 (55.6%)
Training set47 (40–70)265 (51.0%)
* Median (IQ range).
Table 3. Average metric values for each class for both models.
Table 3. Average metric values for each class for both models.
Model-8Model-5p-Value
1. Chondro-osseous border (IoU)0.6320.624<0.001
2. Femoral head (IoU)0.9160.9150.26
3. Labrum (IoU)0.6210.625<0.01
4. Cartilaginous roof (IoU)0.6660.672<0.001
5. Bony roof (IoU)0.7160.7250.89
6. Bony rim (Euclidean distance) [pixels]4.8-
7. Lower limb (Euclidean distance) [pixels]4.5-
8. Baseline (deflection angle) [degrees]1.7-
Table 4. Success rate for each class for both models.
Table 4. Success rate for each class for both models.
Model-8Model-5
Chondro-osseous border85.88%84.56%
Femoral head100.00%100.00%
Labrum82.41%85.76%
Cartilaginous roof97.79%98.50%
Bony roof97.85%98.03%
Table 5. Dice similarity coefficient for each class for both models.
Table 5. Dice similarity coefficient for each class for both models.
Model-8Model-5
Chondro-osseous border0.7740.769
Femoral head0.9560.956
Labrum0.7670.769
Cartilaginous roof0.7990.804
Bony roof0.8340.841
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Pulik, Ł.; Czech, P.; Kaliszewska, J.; Mulewicz, B.; Pykosz, M.; Wiszniewska, J.; Łęgosz, P. Artificial Intelligence Algorithm Supporting the Diagnosis of Developmental Dysplasia of the Hip: Automated Ultrasound Image Segmentation. J. Clin. Med. 2025, 14, 6332. https://doi.org/10.3390/jcm14176332

AMA Style

Pulik Ł, Czech P, Kaliszewska J, Mulewicz B, Pykosz M, Wiszniewska J, Łęgosz P. Artificial Intelligence Algorithm Supporting the Diagnosis of Developmental Dysplasia of the Hip: Automated Ultrasound Image Segmentation. Journal of Clinical Medicine. 2025; 14(17):6332. https://doi.org/10.3390/jcm14176332

Chicago/Turabian Style

Pulik, Łukasz, Paweł Czech, Jadwiga Kaliszewska, Bartłomiej Mulewicz, Maciej Pykosz, Joanna Wiszniewska, and Paweł Łęgosz. 2025. "Artificial Intelligence Algorithm Supporting the Diagnosis of Developmental Dysplasia of the Hip: Automated Ultrasound Image Segmentation" Journal of Clinical Medicine 14, no. 17: 6332. https://doi.org/10.3390/jcm14176332

APA Style

Pulik, Ł., Czech, P., Kaliszewska, J., Mulewicz, B., Pykosz, M., Wiszniewska, J., & Łęgosz, P. (2025). Artificial Intelligence Algorithm Supporting the Diagnosis of Developmental Dysplasia of the Hip: Automated Ultrasound Image Segmentation. Journal of Clinical Medicine, 14(17), 6332. https://doi.org/10.3390/jcm14176332

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop