1. Introduction
Adolescent idiopathic scoliosis (AIS) is a characterized by a deformation of the spine in 3-D [
1]. In AIS, observations of the illiac crests, namely, the Risser grade, are routinely used to assess the skeletal maturity [
2]. This information is highly relevant to assess the peak growth rate velocity, the growth rate and the final adult height [
3]. This information is crucial for the management of AIS and guiding treatment [
4].
Two common methods for assessing bone age are the Greulich-Pyle (GP) [
5] and Tanner-Whitehouse (TW) [
6] methods. In the GP method, doctors compare X-rays of the hand and wrist with a reference book of images to estimate bone age. The TW method is more detailed, in which clinicians score each bone in the hand and wrist, adding up the scores to determine bone age. For AIS patients, the Sanders classification is often used. This method is simpler than the TW method and was designed to be quick, reliable, and helpful for tracking the type and severity of the spinal curve [
7]. Some examples of wrist radiographs of AIS patients are shown in
Figure 1.
Lee et al. [
8] evaluated 3000 hand images annotated with key feature points, enabling the precise selection of relevant regions for age estimation. Distinct regions of interest (ROIs) were defined, including small areas of carpal and metacarpal bones and larger sections encompassing phalanges. Histogram equalization was applied to minimize irrelevant intensity variations in cropped images. Addressing gender-based growth rate differences, separate models were developed. Notably, the research embraced all age ranges, including scarce data from infancy and early childhood, often overlooked due to morphological differences. By training various deep learning architectures with different ROI definitions, the study achieved a minimum mean absolute difference error of 8.890 months on a test set of 400 images. This preliminary investigation paves the way for future research into unexplored alternative approaches.
Kaddioui et al. [
9], retrospectively collected and manually graded 1830 posteroanterior radiographs of adolescent patients with AIS using the United States Risser staging system. The radiographs underwent preprocessing and were cropped to focus on the pelvic region. A convolutional neural network (CNN) was trained to automate Risser classification, achieving accuracy comparable to human graders. The network’s performance was validated against interobserver variability using the Fleiss measure. Due to limited radiograph availability, transfer learning was employed with the VGG16 network. The agreement among six observers was moderate, with a coefficient of 0.65 and 74.5% concordance. The automated method achieved substantial agreement (=0.72) and an overall accuracy of 78.0% compared to ground truth. This research demonstrates the potential of deep learning in automating AIS grading, even with a limited dataset, and establishes a promising foundation for improving scoliosis assessment in clinical practice. Magnide et al. [
10] have used a pre-trained ResNet101 model, with a SVM classifier to classify different region of interests on pancorporeal radiographs of AIS patients (pelvis, humeral heads, femoral heads) at 5 different visits. Visualisation of class activation maps (CAM) using gradient-weighted class activation maps (Grad-CAM) was used to outline the different ossification centers and validation for follow up. Unfortunately, the posture of AIS patient does not allow a good visualisation of the hand and a second radiograph of the hand/wrist was often required to assess precisely the bone age. Recently, we have introduced a new posture in pancorporeal radiographs, to include the wrist and hand.
Pan et al. [
11] proposed a comprehensive evaluation of bone age assessment methods using deep learning models and expert radiologists. The study focuses on a Convolutional Neural Network (CNN) named TDL-BAAM, trained on 15,129 pediatric trauma hand radiographs from Children’s Hospital of New York. CNN’s predictive ability was compared against a GP-based deep learning model (GPDL-BAAM) and two pediatric radiologists using the GP method. An independent test set of 214 trauma hand radiographs from Hasbro Children’s Hospital was used for validation. The TDL-BAAM achieved a mean absolute error (MAE) of 11.1 months, outperforming GPDL-BAAM (12.9 months) and the radiologists (14.6 and 16.0 months). TDL-BAAM’s predictions were within 24 months of chronological age for 95.3% of cases, compared to 91.6% for GPDL- BAAM and lower percentages for radiologists. High concordance (Intraclass Correlation Correlation Coefficient ICC: 0.93) existed between all methods and chronological age. The deep learning models showed a systematic bias, leaning towards overestimating age for younger children, unlike radiologists who exhibited consistent biases. This study underscores the potential of deep learning for accurate bone age assessment, with TDL-BAAM demonstrating notable performance improvements over traditional methods and human experts.
In recent years, the recent advances in artificial intelligence, particularly deep learning, have found to be promising for medical image analysis [
12,
13]. Convolutional neural networks (CNNs) might contribute to reducing the reading time compared to methods. Despite this progress, the transition from research to clinical application necessitates rigorous validation and adaptation to real-world scenarios. The application of deep learning models in medical contexts demands external validation studies employing diverse datasets to ensure generalizability and robustness.
The goal of this study is to present a new bone age classification approach based on deep neural networks and on class activation maps for wrist radiographs from EOS radiographs. We evaluated our results on the Radiological Society of North America (RSNA) Bone Age dataset and on clinical EOS radiograph AIS patients.
2. Methodology
2.1. Dataset
Our model was trained on the RSNA Bone Age dataset. The RSNA Bone Age dataset [
14] is a freely accessible set of data created to support the development and testing of automated skeletal age estimation techniques. It comprises 12,000 hand radiographs at different ages, from infants to late teens as depicted in
Table 1. Both male and female participants are represented in the dataset. The radiographs were acquired from several hospitals and imaging facilities. Some sample images from the dataset are shown in
Figure 2.
The proposed model was evaluated on a cohort of 257 clinical AIS patients. First, the region of interest of the hand and wrist was manually extracted from full-standing EOS radiographs using GNU image processing software (GIMP) which is free and open-source. Then, the bone age of AIS patient was evaluated by using BoneXpert (Visiana ApS, Hørsholm, DEN) [
15]. BoneXpert, employs advanced active appearance models (AAMs) and machine learning techniques to automate bone age estimation [
16], with a reported accuracy of 95%. By reconstructing bone borders, computing intrinsic bone ages based on shape, intensity, and texture scores, and subsequently transforming them into GP or TW bone ages.
The skeletal age of the patient was obtained by a corresponding skeletal age label that is present with every hand radiograph in the dataset. The maturation of certain bones is evaluated, and their comparison to accepted reference atlases yields the bone age. The dataset includes information about the patient’s age, gender, and hand side (left or right), in addition to the hand radiographs and bone age labels. The bone age distribution of various patients are shown in
Figure 3.
In terms of dataset utilization for model training and testing, a split ratio of 80:20 is employed. This allocation ensures that 80% of the data is used for training the model, allowing it to learn and adapt to various patterns and nuances present in the radiographs. The remaining 20% of the data is reserved for testing, providing a robust means to evaluate the model’s effectiveness and accuracy in bone age estimation.
2.2. Initial Experiments
The initial exploration in bone age estimation involved conducting experiments with three transfer learning models: Visual Geometry Group (VGG16 and VGG19) (University of Oxford, Department of Engineering Science, Oxford, UK) and Inception (Google AI, Google LLC., Mountain View, CA, USA). The relatively high MAE values for each model underscored their limitations. The experiments with VGG16, VGG19, and Inception models were conducted under uniform hyperparameters: a batch size of 32, the Rectified Linear Unit (ReLU) activation function, and the Adam optimizer. VGG16 consists of 16 layers, VGG19 has 19 layers, and Inception involves a more intricate architecture with multiple convolution layers. Despite this, their performance was suboptimal, as indicated by high MAE values. This could be attributed to their architectures being less suited for the specific nuances of bone age imaging, possibly lacking in capturing fine-grained details critical for accurate age estimation. These preliminary trials were essential in guiding the direction of our research.
2.3. Experimental Design
The bone age estimation model uses X-ray images of wrists as input to the proposed deep learning (DL) model. The model employs a fine-tuned transfer learning approach with DenseNet201 (Linux Foundation, San Francisco, CA, USA) for maximum accuracy and minimal MAE and leverages DenseNet201’s pre-trained features for efficient feature extraction. DenseNet201 is fine-tuned by taking it as a base layer and then integrating our own uniquely designed CNN layers. This approach significantly reduces the computational cost and training time required while still delivering good accuracy. The proposed experimental bone age estimation neural network architecture is shown in
Figure 4.
2.4. Network Architecture
Each image from the RSNA bone age dataset undergoes preprocessing, including resizing to 299 × 299, to conserve computational resources. This resizing step was crucial for speeding up the processing and minimizing memory consumption. The first layer of our architecture consists of “Densenet201” as a functional transfer learning layer that takes the input image and performs convolutional operations to extract features. The architecture of DenseNet201 consists of multiple densely connected blocks, each containing several layers. Unlike traditional CNNs, DenseNet201 employs skip connections, or shortcuts, that connect every layer to every other layer within the same block. This dense connectivity pattern not only facilitates feature reuse but also addresses the vanishing gradient problem commonly encountered in deep networks. The Densenet201 instantiation notably omits the fully connected layer and integrates pre-trained weights from the ImageNet (Stanford Vision Lab, Stanford University, Stanford, CA, USA) database, a standard practice in transfer learning to leverage pre-learned feature mappings. The “Densenet201” layer gives an output shape of (None, 9, 9, 1920), indicating that the input image is transformed into a feature map with a size of 9 × 9 and 1920 channels.
Once the DenseNet201 architecture has processed the input, its output serves as the input for our custom-designed CNN layers. Our CNN architecture is designed to better recognize and understand the detailed patterns in the X-ray images.
Figure 5 shows the detailed description of the DL model used for bone age prediction.
Subsequent to the integration of DenseNet201, the proposed custom CNN employs a layer. This layer effectively reduces the dimensions of the input and concentrates on the important features required for bone age estimation. This reduction is followed by a layer, which transforms the condensed feature maps into a linear array. The model further consists of a sequence of three dense layers. The first and second of these layers comprise 64 and 32 neurons, respectively, and employ the ‘relu’ activation function. This function is instrumental in introducing non-linear characteristics to the model, enabling it to decipher more intricate patterns in the data. The final layer in this sequence is a single neuron layer utilizing a ‘linear’ activation function, implying that the model’s output is a singular continuous variable such as the bone age.
Regarding optimization techniques, multiple optimizers are defined: stochastic Gradient descent (SGD), nesterov-accelerated adaptive moment estimation (Nadam), and Adamax. However, the compilation of the proposed custom CNN is exclusively with the Adamax optimizer. This optimizer is configured with a learning rate of 0.001, and the model employs a mean squared error (MSE) loss function, which aligns with regression-based objectives. The primary metric for evaluating model performance is articulated as ‘’, representing mean absolute error in months, a metric particularly pertinent in applications of precise continuous bone age value predictions. In terms of the number of parameters, the model has 18,447,041 parameters with a total batch size of 32. Among them, 18,217,985 parameters are trainable, meaning they are updated during the training process to optimize the model’s performance. The remaining 229,056 parameters are non-trainable, representing fixed weights that are not updated.
2.5. Model Explainability: Score-CAM
CAMs also called heatmaps, are a way to understand how CNNs make their predictions. In this study, we used a method called score-weighted class activation maps (Score-CAM), which measures the importance of different parts of an image by looking at how input features affect the output. Unlike traditional methods that rely on gradients (a kind of sensitivity measurement), Score-CAM takes a broader view of the input’s impact. It [
17] is also useful for identifying issues in the model, such as why it might make an incorrect prediction or if there are biases in the dataset. This makes it a powerful tool for improving and understanding deep learning models.
The reason for choosing Score-CAM [
18] lies in the fact that it typically outperforms other methods across a wide range of metrics, with the exception of complexity. Furthermore, in [
18], it’s worth mentioning that Score-CAM shows strong performance with VGG and ResNet architectures. Conversely, Grad-CAM [
19] scores the highest in the ADCC score with ResNeXt models.
2.6. Evaluation Metrics
Mean absolute error (
): MAE [
20], generally expressed in months for bone age, is a measurement of the average size of errors in a set of predictions. It is used to evaluate the efficacy of a regression model and is quantified as the average absolute difference between the predicted values and the actual values. The MAE loss function formula:
where:
n is the number of observations in the dataset,
is the true value,
is the predicted value
“MAE months” calculates MAE in months, thereby quantifying the discrepancy between anticipated bone ages and factual bone ages. This metric has significant medical relevance, facilitating an intelligible measure of the model’s precision in bone age estimation. This function, characterized by two inputs, normalized predicted bone ages and normalized actual bone ages transforms these values back to their original scale using calculated mean and standard deviation from the training dataset. This transformation yields an MAE in months, a temporal dimension of considerable clinical significance. A decreased MAE signifies better alignment between projections and factual bone ages.
3. Results
The proposed neural network, trained for 35 epochs, proposes a solution of automating bone age assessment using machine learning. Furthermore, we explored the benefits of using an automated boneage prediction method using DL, which leads to improved MAE, accuracy, efficiency, and decision making for support in managing scoliosis patients.
To assess the effectiveness of various models in bone age estimation, initially several experiments with VGG16, VGG19, and Inception, were done while ensuring consistency in hyperparameters across all trials. A detailed evaluation of the performance of each model was done, where VGG16 showed a MAE of 11.35 months, VGG19 had an MAE of 14.34 months, and Inception yielded an MAE of 32.5 months.
Figure 6 presents a detailed graphical comparison of the performance metrics in these different models.
In a scatter plot as shown in
Figure 6, the model’s accuracy is inferred by how closely the data points cluster around the line representing the actual age. The VGG16 model on the top left plots a noticeable improvement in prediction accuracy. The points are more tightly clustered around the line, especially in the mid-range of ages. However, some discrepancies remain for younger and older ages. Moving to the top right VGG19 plot, it depicts a model with significant variance, as the data points are quite spread out from the actual age line. This model tends to underpredict ages for younger subjects and overpredict for older ones, which is suggested by the curvature of the data points away from the actual age line. The bottom left Inception plot displays the weakest predictive performance among the four. The data points show a considerable spread, indicating high MAE, and there is a consistent trend where the model underpredicts the actual age across the full age range. In contrast, the bottom right proposed model plot showcases the most accurate model. The predictions are closely aligned with the actual ages, with data points densely packed around the line, showing low MAE values and minimal bias. Overall, the proposed model is the most precise, while the Inception model significantly underperforms. The resulting MAE of 4.87 months and an
value of 0.975 achieved by our custom network demonstrates the effectiveness of the applied approach and justifies the decision to invest in the development of a specialized model.
The predicted bone age values generated by our custom neural network were compared with the actual bone age labels for the hand radiographs in
Figure 7. It indicates that the proposed model generally predicts with a high degree of accuracy, as evident by the close proximity of the predicted and actual ages in several instances, such as a predicted age of 7.1 years being near an actual age of 7.2 years. However, there are some instances where the proposed model’s predictions deviate from the actual predictions, as seen in an outlier where the model predicted 13.0 years compared to the actual age of 11.4 years, this can be attributed to various factors related to the inconsistencies in image quality such as contrast, brightness, and noise levels. Furthermore, the model might struggle with atypical bone structures or rare pathologies that are underrepresented in the training set, leading to inaccurate predictions for such cases.
The results show a similarity between the predicted and actual bone ages. The small difference between the predicted and actual values further validates the effectiveness of our approach in accurately assessing bone age in hand radiographs.
Figure 8 illustrates the model’s loss graph, depicting its training progress and validation process, and also displays the MAE Month graph, measuring the training and validation
values. Using automation and advanced algorithms, our approach enhances the diagnostic process and facilitates personalized treatment planning, ultimately resulting in better patient outcomes.
Table 2 presents a comparison of the performance of the proposed bone age CNN model with different models from the literature.
In addition, we utilized the Score-CAM model as saliency maps of Score-CAM are more focused than Grad-CAM [
17]. Score-CAM was used to generate visual maps that evaluate the performance of our proposed method, which visually represents the areas in the wrist bone images that contribute most significantly to accurate bone age in months. The results of the feature activation heat map analysis are depicted in
Figure 9, illustrating the effectiveness of our approach in identifying and highlighting the key bone regions. These heat maps provide valuable insights into the regions that play a crucial role in determining bone age and aid in understanding the decision-making process of our deep neural network. This process enhances the robustness of our method without requiring additional segmentation steps, reducing computational complexity and improving efficiency. The Score-CAM image provides a visual representation of how the neural network processes data for bone age estimation in AIS patients. The areas where the network concentrates its attention are distinctly marked in blue, indicating regions of high focus. In contrast, the areas shown in red represent regions where the network’s focus is significantly lower.
The
was used as the evaluation metric, offering a quantitative measure of the predictive models’ performance in estimating bone age accurately. The results of the proposed method as compared to the performance of BoneXpert are shown in
Figure 10. The bar chart compares the predicted bone age from two different models: the proposed model and BoneXpert, across various age groups. Both models yield similar results, with slight variations in predictions. The similarity between the
values of both the BoneXpert system and the proposed research model underlines their comparable performance in predicting bone age.
4. Discussion
The primary question this research aims to address is whether a custom CNN can reliably and accurately assess bone age in AIS patients using hand radiographs. This is particularly important for evaluating skeletal maturity, which is a critical factor in managing AIS progression and treatment planning. The study also examines the potential of explainability tools, such as Score-CAM, to improve the transparency of the model in clinical settings. The proposed model demonstrates promising results in predicting bone age, closely matching BoneXpert in several instances. For example, the proposed model’s prediction is close to BoneXpert’s. Furthermore, when the proposed model predicts lower than BoneXpert, like the 15.95 versus 14.62 years, it suggests the model could be more sensitive in detecting earlier maturation stages, which can be crucial for early intervention strategies. Even though this is a preliminary study, the results indicate a good correlation with the BoneXpert model.
The paper’s methodology section outlines the development of a custom neural network (NN) architecture based on DenseNet201, leveraging transfer learning to achieve low MAE for AIS patients is novel. Unlike many general bone age assessment studies, this work specifically addresses AIS, a niche yet clinically significant subgroup. This approach not only reduces computational complexity but also yields a MAE of 4.87 months, surpassing other pre-trained models like VGG16 and VGG19, demonstrating the effectiveness of domain-specific model development. Furthermore, the paper introduces Score-CAM as an explainability tool, shedding light on the regions of interest within hand radiographs that contribute most significantly to accurate bone age estimation. This visual explanation enhances the trust and transparency of the DNN model, making it more acceptable in clinical practice. Introducing Score-CAM to visualize the regions of interest in the radiographs highlights a unique aspect of this study, ensuring clinical relevance and interpretability.
The study compares the proposed DNN’s performance with the BoneXpert system, a leading approach in the field. The use of automated techniques to evaluate the bone age might be highly relevant in the study of spinal deformities. Our results reveal similar MAE values, showcasing the potential of the proposed DNN to be a reliable and efficient tool for bone age prediction. This convergence in performance signifies the progress in automating bone age assessment, reducing inter-rater variability, and expediting clinical decision-making. The ability of the proposed method to extract key areas of the bone in an explainable way is a significant advantage. By gaining insight into which regions are influential in the classification process, clinicians and radiologists can better understand the algorithm’s decision and have greater confidence in its results. The explainability aspect is particularly important in medical applications as it enables trust, transparency, and acceptance of the automated system. This study addresses a critical gap in the field: the need for automated and interpretable tools for bone age assessment in AIS patients. Current approaches often rely on subjective manual assessments or black-box AI models with limited clinical adoption due to a lack of transparency. By combining automation with interpretability, the proposed approach aims to bridge the gap between technological advances and clinical usability.
The proposed methodology also has some limitations which could be improved in the future works. Our method requires to outline manually the region of interest of the hand from the pancorporal radiograph. This step can be automated using a recent object detection model such as Masked R-CNN, but it remains to be evaluated. The proposed bone age estimation model may exhibit biases due to data imbalances and lacks representation of diverse boneage demographics. Its accuracy heavily depends on the input X-ray image quality and varies with equipment differences. Additionally, the model’s reliability in rare conditions or anomalies affecting bone development, not well-represented in the training data, is uncertain. Furthermore, the use of DenseNet201, known for its densely connected architecture, along with additional custom layers, could lead to a computationally intensive model. This might require significant computational resources for training and inference, potentially limiting its use in environments with limited processing capabilities. Furthermore, researchers could explore the integration of advanced object detection models, such as Masked R-CNN or YOLO, to automate the extraction of the hand region from pancorporal radiographs. This will not only streamline the workflow, but also reduce potential human error and improve reproducibility. Future models should be trained with additional data that include rare anomalies that affect bone development. Collaboration with specialized medical institutions could provide access to such datasets, enhancing applicability of the model in less common clinical scenarios.
5. Conclusions
This study presents a deep neural network-based approach for pediatric bone age assessment, demonstrating its potential for the field of pediatric radiology in general, but also applied to AIS patients. The custom DNN architecture, combined with transfer learning and Score-CAM for explainability, results in accurate and transparent bone age predictions. To further enhance the model’s accuracy and robustness, collecting and annotating a more extensive and diverse dataset of pediatric hand radiographs, encompassing various age groups, ethnicities, and clinical conditions, would be beneficial. Data augmentation techniques can also be employed to artificially increase the dataset size. At the moment, the identification of the region of interest of the hand and wrist is manual, but future work will involve the automatic identification of the region of interest and the evaluation of BoneXpert on a larger amount of AIS patients. The adoption of deep neural networks in bone age assessment holds promise for better monitoring of growth progression in pediatric populations.
Author Contributions
Conceptualization, I.A., B.Z.H. and L.D.; data curation, I.A. and L.D.; methodology, I.A. and B.Z.H.; project administration, L.D.; resources, J.J., S.B., M.R.-B., S.P., G.G., H.L. and L.D.; software, I.A. and B.Z.H.; supervision, J.J., S.B., M.R.-B., S.P., G.G., H.L. and L.D.; validation, J.J., S.B., M.R.-B., S.P., G.G., H.L. and L.D.; writing—original draft preparation, I.A. and B.Z.H.; writing—review and editing, I.A., B.Z.H., J.J., S.B., M.R.-B., S.P., G.G., H.L. and L.D. All authors have read and agreed to the published version of the manuscript.
Funding
This research was funded by Mitacs Globalink program and the Chaire de recherche en science du mouvement de l’Université de Montréal.
Institutional Review Board Statement
This project was approved by the Institutional Review Board 379 of CHU Sainte-Justine (IRB# 2018-1377, approval date 20 June 2024).
Informed Consent Statement
Informed consent was waived for conducting this study.
Data Availability Statement
Conflicts of Interest
The authors declare no conflicts of interest.
Abbreviations
The following abbreviations are used in this manuscript:
AIS | Adolescent Idiopathic Scoliosis |
CNN | Convolutional Neural Network |
CAM | Class Activation Map |
GP | Greulich-Pyle |
TW | Tanner-Whitehouse |
RSNA | Radiological Society of North America |
MAE | Mean Absolute Error |
ROI | Region of Interest |
GPDL-BAAM | Greulich-Pyle Deep Learning Bone Age Assessment Model |
TDL-BAAM | Tanner-Whitehouse Deep Learning Bone Age Assessment Model |
ICC | Intraclass Correlation Coefficient |
VGG | Visual Geometry Group |
Inception | Deep Learning Model Architecture (GoogleNet) |
DenseNet201 | Dense Convolutional Network with 201 Layers |
ReLU | Rectified Linear Unit |
MSE | Mean Squared Error |
SVM | Support Vector Machine |
Score-CAM | Score-Weighted Class Activation Map |
Grad-CAM | Gradient-Weighted Class Activation Map |
EOS | End of Scoliosis |
AAM | Active Appearance Model |
BoneXpert | Commercial Bone Age Grading Software |
GIMP | GNU Image Manipulation Program |
References
- Kadoury, S.; Mandel, W.; Roy-Beaudry, M.; Nault, M.; Parent, S. 3D Morphology Prediction of Progressive Spinal Deformities from Probabilistic Modeling of Discriminant Manifolds. IEEE Trans. Med. Imaging 2017, 36, 1194–1204. [Google Scholar] [CrossRef]
- Heegaard, M.; Tøndevold, N.; Dahl, B.; Andersen, T.B.; Gehrchen, M.; Ohrt-Nissen, S. Does Risser stage accurately predict the risk of curve progression in patients with adolescent idiopathic scoliosis treated with night-time bracing? Eur. Spine J. 2023, 32, 3077–3083. [Google Scholar] [CrossRef] [PubMed]
- Cavallo, F.; Angelika, M.; Chiarelli, F.; Giannini, C. Evaluation of Bone Age in Children: A Mini-Review. Front. Pediatr. 2021, 9, 580314. [Google Scholar] [CrossRef]
- Sanders, J.; Khoury, J.; Kishan, S.; Browne, R.; Mooney, J.; Arnold, K.; McConnell, S.; Bauman, J.; Finegold, D. Predicting Scoliosis Progression from Skeletal Maturity: A Simplified Classification During Adolescence. J. Bone Jt. Surg. Am. 2008, 90, 540–553. [Google Scholar] [CrossRef]
- Tsehay, B.; Afework, M.; Mesifin, M. Assessment of Reliability of Greulich and Pyle (GP) Method for Determination of Age of Children at Debre Markos Referral Hospital, East Gojjam Zone. Ethiop. J. Health Sci. 2017, 27, 631. [Google Scholar] [CrossRef]
- Prokop-Piotrkowska, M.; Marszałek-Dziuba, K.; Moszczyńska, E.; Szalecki, M.; Jurkiewicz, E. Traditional and New Methods of Bone Age Assessment-An Overview. J. Clin. Res. Pediatr. Endocrinol. 2020, 13, 251–262. [Google Scholar] [CrossRef] [PubMed]
- Suggala, S.; Minopoli, A.L.; Shahid, A.H.; Menger, R.P. Transitional and spectrum-based care for idiopathic scoliosis. Neurosurg. Focus 2024, 57, E11. [Google Scholar] [CrossRef] [PubMed]
- Lee, J.H.; Kim, Y.J.; Kim, K.G. Bone age estimation using deep learning and hand X-ray images. Biomed. Eng. Lett. 2020, 10, 323–331. [Google Scholar] [CrossRef] [PubMed]
- Kaddioui, H.; Duong, L.; Joncas, J.; Bellefleur, C.; Nahle, I.; Chémaly, O.; Nault, M.L.; Parent, S.; Grimard, G.; Labelle, H. Convolutional Neural Networks for Automatic Risser Stage Assessment. Radiol. Artif. Intell. 2020, 2, e180063. [Google Scholar] [CrossRef] [PubMed]
- Magnide, E.; Tchaha, G.; Joncas, J.; Bellefleur, C.; Barchi, S.; Roy-Beaudry, M.; Parent, S.; Grimard, G.; Labelle, H.; Duong, L. Automatic bone maturity grading from EOS radiographs in Adolescent Idiopathic Scoliosis. Comput. Biol. Med. 2021, 136, 104681. [Google Scholar] [CrossRef] [PubMed]
- Pan, I.; Baird, G.; Mutasa, S.; Merck, D.; Ruzal-Shapiro, C.; Swenson, D.; Ayyala, R. Rethinking Greulich and Pyle: A Deep Learning Approach to Pediatric Bone Age Assessment Using Pediatric Trauma Hand Radiographs. Radiol. Artif. Intell. 2020, 2, e190198. [Google Scholar] [CrossRef]
- Nadeem, W.; Goh, H.; Ali, A.; Hussain, M.; Khan, M.; Ponnusamy, V. Bone Age Assessment Empowered with Deep Learning: A Survey, Open Research Challenges and Future Directions. Diagnostics 2020, 10, 781. [Google Scholar] [CrossRef] [PubMed]
- Hussain, B.Z.; Andleeb, I.; Ansari, M.S.; Kanwal, N. Lightweight Deep Learning Model for Automated COVID-19 Diagnosis from CXR Images. In Proceedings of the 2021 IEEE International Conference on Computing (ICOCO), Kuala Lumpur, Malaysia, 17–19 November 2021; pp. 218–223. [Google Scholar] [CrossRef]
- Halabi, S.S.; Prevedello, L.M.; Kalpathy-Cramer, J.; Mamonov, A.B.; Bilbily, A.; Cicero, M.; Pan, I.; Pereira, L.A.; Sousa, R.T.; Abdala, N.; et al. The RSNA Pediatric Bone Age Machine Learning Challenge. Radiology 2019, 290, 498–503. [Google Scholar] [CrossRef] [PubMed]
- Bonexpert. Bonexpert—Bone Health Assessment. 2009. Available online: https://bonexpert.com/ (accessed on 18 January 2025).
- Thodberg, H.H.; Kreiborg, S.; Juul, A.; Pedersen, K.D. The BoneXpert Method for Automated Determination of Skeletal Maturity. IEEE Trans. Med. Imaging 2009, 28, 52–66. [Google Scholar] [CrossRef] [PubMed]
- Wang, H.; Wang, Z.; Du, M.; Yang, F.; Zhang, Z.; Ding, S.; Mardziel, P.; Hu, X. Score-CAM: Score-Weighted Visual Explanations for Convolutional Neural Networks. arXiv 2020, arXiv:1910.01279. [Google Scholar]
- Poppi, S.; Cornia, M.; Baraldi, L.; Cucchiara, R. Revisiting the evaluation of class activation mapping for explainability: A novel metric and experimental analysis. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 2299–2304. [Google Scholar]
- Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-cam: Visual explanations from deep networks via gradient-based localization. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 618–626. [Google Scholar]
- Wang, W.; Lu, Y. Analysis of the Mean Absolute Error (MAE) and the Root Mean Square Error (RMSE) in Assessing Rounding Model. IOP Conf. Ser. Mater. Sci. Eng. 2018, 324, 012049. [Google Scholar] [CrossRef]
- Kim, P.H.; Yoon, H.M.; Kim, J.R.; Hwang, J.Y.; Choi, J.H.; Hwang, J.; Lee, J.; Sung, J.; Jung, K.H.; Bae, B.; et al. Bone Age Assessment Using Artificial Intelligence in Korean Pediatric Population: A Comparison of Deep-Learning Models Trained with Healthy Chronological and Greulich-Pyle Ages as Labels. Korean J. Radiol. 2023, 24, 1151–1163. [Google Scholar] [CrossRef] [PubMed]
- Nam, Y. Bone Age Estimation Using Machine Learning Approach: An Assessment Using Hand and Wrist Bones of South African Children. Ph.D. Thesis, University of Cape Town, Cape Town, South Africa, 2023. [Google Scholar]
- Toka, D.; Kırcı, M. Determination of DL-Based Bone Age Assessment. In Proceedings of the 2023 14th International Conference on Electrical and Electronics Engineering (ELECO), Bursa, Turkey, 30 November–2 December 2023; pp. 1–5. [Google Scholar] [CrossRef]
- Guo, L.; Wang, J.; Teng, J.; Chen, Y. Bone Age Assessment Based on Deep Convolutional Features and Fast Extreme Learning Machine Algorithm. Front. Energy Res. 2022, 9, 813650. [Google Scholar] [CrossRef]
- Poojary, N.; Pokhare, P.; Poojary, P.; Raghavani, C.; Khanapuri, J. A Novel Approach for Bone Age Assessment using Deep Learning. Int. J. Sci. Res. Comput. Sci. Eng. Inf. Technol. 2021, 7, 67–75. [Google Scholar] [CrossRef]
- Li, S.; Liu, B.; Li, S.; Zhu, X.; Yan, Y.; Zhang, D. A deep learning-based computer-aided diagnosis method of X-ray images for bone age assessment. Complex Intell. Syst. 2021, 8, 1929–1939. [Google Scholar] [CrossRef] [PubMed]
| Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).