Next Article in Journal
Staging of Liver Fibrosis Based on Energy Valley Optimization Multiple Stacking (EVO-MS) Model
Previous Article in Journal
Artificial Intelligence Support for Informal Patient Caregivers: A Systematic Review
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Application of Artificial Intelligence Methods on Osteoporosis Classification with Radiographs—A Systematic Review

by
Ren Wei Liu
1,*,†,
Wilson Ong
1,*,†,
Andrew Makmur
1,2,
Naresh Kumar
3,
Xi Zhen Low
1,2,
Ge Shuliang
1,2,
Tan Yi Liang
1,2,
Dominic Fong Kuan Ting
1,2,
Jiong Hao Tan
3 and
James Thomas Patrick Decourcy Hallinan
1,2
1
Department of Diagnostic Imaging, National University Hospital, 5 Lower Kent Ridge Road, Singapore 119074, Singapore
2
Department of Diagnostic Radiology, Yong Loo Lin School of Medicine, National University of Singapore, 10 Medical Drive, Singapore 117597, Singapore
3
University Spine Centre, Department of Orthopaedic Surgery, National University Health System, 1E, Lower Kent Ridge Road, Singapore 119228, Singapore
*
Authors to whom correspondence should be addressed.
These authors contribute equally to this work.
Bioengineering 2024, 11(5), 484; https://doi.org/10.3390/bioengineering11050484
Submission received: 23 March 2024 / Revised: 24 April 2024 / Accepted: 7 May 2024 / Published: 12 May 2024

Abstract

:

Simple Summary

Osteoporosis is a major global health problem with substantial economic and psychosocial repercussions. Underdiagnosis of osteoporosis is prevalent. The dual-energy absorptiometry (DEXA) scan is the gold standard for bone mineral density (BMD) measurement but its accessibility is limited. Radiographs are ubiquitous in healthcare and represent a promising avenue for opportunistic osteoporosis screening. Historically, this has been associated with high labor and time costs but several recent studies have demonstrated that BMD can be estimated from radiographs in a cost-effective manner utilizing deep learning techniques. This review aims to summarize the existing evidence supporting the utility of artificial intelligence (AI) methods for osteoporosis classification using radiographs.

Abstract

Osteoporosis is a complex endocrine disease characterized by a decline in bone mass and microstructural integrity. It constitutes a major global health problem. Recent progress in the field of artificial intelligence (AI) has opened new avenues for the effective diagnosis of osteoporosis via radiographs. This review investigates the application of AI classification of osteoporosis in radiographs. A comprehensive exploration of electronic repositories (ClinicalTrials.gov, Web of Science, PubMed, MEDLINE) was carried out in adherence to the Preferred Reporting Items for Systematic Reviews and Meta-Analyses 2020 statement (PRISMA). A collection of 31 articles was extracted from these repositories and their significant outcomes were consolidated and outlined. This encompassed insights into anatomical regions, the specific machine learning methods employed, the effectiveness in predicting BMD, and categorizing osteoporosis. Through analyzing the respective studies, we evaluated the effectiveness and limitations of AI osteoporosis classification in radiographs. The pooled reported accuracy, sensitivity, and specificity of osteoporosis classification ranges from 66.1% to 97.9%, 67.4% to 100.0%, and 60.0% to 97.5% respectively. This review underscores the potential of AI osteoporosis classification and offers valuable insights for future research endeavors, which should focus on addressing the challenges in technical and clinical integration to facilitate practical implementation of this technology.

Graphical Abstract

1. Introduction

1.1. Osteoporosis—A Global Health Challenge

Osteoporosis is a complex endocrine disorder affecting bone turnover, marked by a decline in bone mass and microstructural integrity [1]. Although asymptomatic in the early stages, it significantly elevates the risk of fragility fractures, leading to increased morbidity, mortality, and reduced quality of life [2].
The disease constitutes a major global health problem [3,4], affecting one-third of women and one-fifth of men aged 50 and above [5,6,7]. Worldwide, an estimated 200 million women suffer from osteoporosis [8], a figure that is projected to rise with the aging global population and increasing life expectancy. The economic and psychosocial repercussions of fragility fractures are enormous [9,10,11,12], with a 2005 United States study reporting direct costs exceeding USD 17 billion [13] and a 2013 European Union study estimating a total economic burden of EUR 37 billion and loss of 1,180,000 quality-adjusted life years in 2010 [14].
Although cost-effective treatments are readily available [15], underdiagnosis and undertreatment remain prevalent. Importantly, out of female patients suffering their first osteoporotic fracture, only an estimated 10.3% will have had undergone prior bone mineral density (BMD) testing [14,16,17,18].

1.2. Current Diagnostic Methods and Challenges

Dual-energy absorptiometry scans (DEXA) of the lumbar spine and hip, advocated by the World Health Organisation (WHO) since 1987, is a gold standard in BMD measurement. DEXA is non-invasive and cost-effective and remains the most frequently utilized radiologic technique for evaluating bone mass [8,19,20]. It is also validated for use with other tools such as the WHO fracture risk assessment (FRAX) algorithm [20]. However, DEXA has its downsides including limited accessibility, need for strict quality standards, operator dependency [21], and suboptimal screening rates [12,18]. Alternative techniques for measuring BMD such as quantitative computed tomography [22], ultrasound [23], and peripheral DEXA [24] are yet to see widespread clinical use.

1.3. Potential of Machine Learning for Osteoporosis Classification

Radiographs are the most frequently utilized imaging modality in healthcare. The potential for osteoporosis classification via radiographs has been explored since the 1960s. For example, Barnett et al. measured lumbar vertebral concavity along with the femoral and metacarpal cortical thicknesses by hand with a millimeter ruler as an osteoporosis scoring procedure [25,26]. In addition, Garn et al. documented various techniques in the use of densitometry to diagnose osteoporosis [27], including the simultaneous use of metallic phantoms in radiography [28,29]. These initial manual methods were associated with significant manual labor costs and exacting technical processing requirements, limiting the feasibility of large-scale use.
Machine learning, a subset of AI, holds great promise for the automated segmentation and classification of large volumes of medical imaging data. Recent studies employing modern convolutional neural networks and deep learning architectures have demonstrated that BMD can be accurately estimated from radiographs in a cost-effective manner. This review aims to summarize the existing evidence supporting the utility of AI methods for osteoporosis classification using radiographs.

1.4. Research Questions

This review aims to synthesize existing evidence supporting the utility of AI methods for osteoporosis classification using radiographs. Key research questions include:
  • How effective are AI methods in accurately classifying osteoporosis using radiographic data?
  • What are the current technical challenges and practical limitations in osteoporosis diagnosis and classification?
  • What are potential future directions for the use of AI-based classification in osteoporosis management?

2. Materials and Methods

2.1. Literature Review

A query of prominent electronic databases (clinicaltrials.gov, Web of Science, MEDLINE, PubMed) was performed in alignment with the Preferred Reporting Items for Systematic Reviews and Meta-Analyses 2020 statement (PRISMA). The search utilized the following specific search terms: (“neural network*” OR “convolutional neural network*” OR “machine learning” OR “deep learning” OR “AI” OR “artificial intelligence”) AND ((“bone” AND “mineral” AND “density”) OR “BMD”) OR “osteoporosis” OR “absorptiometry” OR “DEXA” OR “osteopaenia” OR “osteopenia” AND (“X-ray” OR (“Radiographs”)). Two reviewers (R.L. and W.O.) screened the resultant articles and shortlisted a number for further examination. The studies were then reviewed and discussed to reach a consensus on their suitability before inclusion. Any disagreements were resolved by a third reviewer (J.T.P.D.H.). The database query was limited to articles published prior to 14 August 2023.

2.2. Screening of Studies and Criteria for Selection

No constraints were specified for the literature and reference search. Articles pertaining to deep learning, artificial intelligence (AI), or deep learning to categorize or predict osteoporosis from radiographs were shortlisted, including those with comparative analysis to conventional DEXA scans when available.

2.3. Extraction of Data and Reporting

Using Microsoft Excel (Microsoft Corporation, Washington, DC, USA), a list of selected papers was compiled. The compiled information encompassed the following:
  • Main clinical use: Classify osteoporosis through the application of machine learning tools (e.g., BMD estimation or bi-variate/tri-variate classification, i.e., normal, osteopenic, and osteoporotic);
  • Research article characteristics: Comprehensive authorship, publication date, and journal or publication name, all written in the English language;
  • Research protocol: Study design, anatomical region, imaging modality, patient demographics, and healthcare context;
  • Machine Learning: Type of machine learning architecture or technique used, need for human supervision or pre-processing.

3. Results

3.1. Search Results

The initial literature review (Figure 1) yielded 607 relevant research articles, which were then screened using the aforementioned criteria. Following the screening, 55 articles were selected, which were further analyzed to determine suitability by two readers. Any discordance was resolved after discussion with a third reader. Upon review, a further 28 publications were excluded as they did not involve AI analysis of radiographs or relied on phantom models rather than patient data. In addition, four articles were added after reviewing the references of the chosen articles. Overall, this resulted in a total of 31 articles (Table 1) for in-depth analysis. The main findings were consolidated and summarized in this systematic review. Due to insufficient data in most studies, formal meta-analysis using 2 × 2 contingency tables was not feasible.

3.2. Model Accuracy in Classification of Osteoporosis

The pooled reported accuracy, sensitivity, and specificity of AI osteoporosis classification ranges from 0.661 to 0.9787, from 0.674 to 1, and from 0.60 to 0.9751, respectively, with AUC values of 0.70 to 0.9987. Most papers used DEXA as the benchmark for comparison. One study assessing knee radiographs used quantitative ultrasound as a reference standard [30] and another examined machine versus human osteoporosis classification based on second metacarpal cortical percentage [31].
Out of the 31 studies, one study examined both hip and spine radiographs, while the other 30 studies examined different anatomical regions as detailed in Table 2. Deep learning models in all anatomical areas showed overall good AUC and accuracy; however, studies on calcaneal and dental radiographs generally yielded higher AUC and accuracy relative to other anatomical regions. For example, Singh et al. performed a study assessing calcaneal radiographs for osteoporosis and noted that the relative paucity of soft tissue around the calcaneum was advantageous, as soft tissue could increase measurement variability. Their study yielded an excellent AUC of 0.9824 and an accuracy of 0.9787 using a support vector machine classifier [32]. In contrast, Cui et al. [33] described the negative impact of soft tissue, bowel gas, and clothing artefacts in their analysis of lumbar spine radiographs and made use of various image processing algorithms to mitigate the problem.
In their respective studies analyzing chest radiographs, Jang et al. [34] noted that it was unclear how incidental findings such as calcified nodules or old fractures may confound osteoporosis classification, while Sato et al. [35] suggested that it may be worthwhile for models to be supplied with clinical information on relevant comorbidities such as fracture history, chronic obstructive pulmonary disease, and rheumatoid arthritis to improve accuracy. The region of assessment can therefore have a significant impact on performance; a model must incorporate various techniques to remain robust in the analysis of different anatomical regions in the clinical setting.

3.3. Study Protocol and Performance Metrics

There is variation in the study protocols and performance metrics used to report model efficacy. For example, some studies performed a binary classification distinguishing osteoporosis and non-osteoporosis and hence relied on metrics such as accuracy. Other authors predicted BMD values for each patient and compared these against the gold standard of DEXA using the correlation co-efficient R. A few of the different performance metrics used are briefly explained below:
  • Accuracy: Accuracy represents the proportion of correctly classified instances among all instances examined. It provides a general measure of model performance but may not be suitable for imbalanced datasets;
  • Sensitivity and Specificity: Sensitivity (true positive rate) measures the proportion of actual positives that are correctly identified by the model, while specificity (true negative rate) measures the proportion of actual negatives that are correctly identified by the model;
  • Area Under the Curve (AUC): AUC refers to the area under the receiver operating characteristic (ROC) curve, which illustrates the trade-off between sensitivity and specificity across various threshold settings. A higher AUC value indicates better discrimination ability of the model;
  • F1 Score: The F1 score is the harmonic mean of precision and recall (sensitivity). It provides a balance between precision (the proportion of true positive predictions among all positive predictions) and recall, making it suitable for imbalanced datasets;
  • Correlation Coefficient (R): The correlation coefficient measures the strength and direction of the linear relationship between two variables. In the context of osteoporosis classification, it reflects the agreement between predicted bone mineral density values (a scalar numerical value) and gold standard measurements obtained from DEXA scans.

3.4. Machine Learning in Medical Imaging

Artificial intelligence is defined in the Merriam-Webster dictionary as “the capability of computer systems or algorithms to imitate intelligent human behaviour”. Machine learning, a subset of artificial intelligence, employs algorithms and statistical techniques to allow computer systems to learn and make informed predictions from data. Several systems have demonstrated diagnostic capabilities comparable to medical professionals in various clinical conditions [36], with many software applications being approved for clinical use [37,38]. Various diagnostic imaging models are available for chest radiographs [39], mammograms [40], and MRI spine analysis [41], with the last showing improved productivity for the reporting radiologist [42]. Radiomics, which refers to the quantitative extraction of various characteristics from a medical image in order to facilitate statistical analysis [43,44], has also been augmented by the use of artificial intelligence.
Figure 2 outlines a general procedure for the development, testing, and deployment of a machine-learning model [45,46,47]. A dataset of medical images would first be collected and pre-processed manually or via automated methods. An appropriate model is then selected and trained using a portion of the dataset (the training set), adjusting internal parameters to improve prediction accuracy. After training, the model is tested on a separate unseen dataset (test set) and fine-tuned to optimize its performance and minimize error.
  • Data Collection: Usually, approval from an ethics committee is necessary before utilizing medical data for the development of a commercial or research AI algorithm. In the case of a prospective study, explicit informed consent is necessary. Medical imaging data are usually collated from a picture archiving and communication system (PACS) environment, requiring collaboration between AI developers and healthcare professionals [48]. Accessing relevant data involves querying, appropriately de-identifying, and securely storing the information. Protected health information must be removed from both the Digital Imaging and Communications in Medicine (DICOM) metadata and the images themselves [49];
  • Image Processing and Segmentation: Segmentation is the process of delineating structures within a medical image, thereby creating structured visual representations from unstructured raw data [47,50]. For example, in tumor segmentation, this could be the process of defining the margins of a tumor [45]. For osteoporosis classification, this could refer to the separation of bone and non-bone structures [51,52];
  • Training and Validation: An appropriate model is selected and trained using a portion of the dataset. Supervised machine learning models are provided with data inputs labeled by human experts, whereas unsupervised models extract salient features from unlabeled data to uncover meaningful relationships within datasets. Models determine how to perform meaningful feature extraction and computation, which involves the evaluation of the image factors that allow a prediction to be made. Image features should be independent of technical factors such as noise, signal, and rotation, as these are common issues in medical images. Machine learning models typically iterate to improve performance with each exposure to the validation set;
  • Testing: The model is then tested on a set of examples to evaluate its diagnostic accuracy and performance [45]. Testing can be performed with internal and external test sets; the former refers to data that come from the same pool as the training set whereas the latter refers to data that have been collected from a different source. Good model performance on external testing sets bolsters confidence in model accuracy, whereas poor performance on external datasets may suggest overfitting [44,53].
Various machine-learning models can be used in medical imaging. In particular, convolutional neural networks (CNN) are a subset of machine learning models, which are frequently applied in medical imaging. CNNs utilize convolution kernels, which move across an input image to create a set of output values that are more suitable for analysis by a neural network [54,55]. Various CNN architectures, for example, LeNet, GoogleNet, AlexNet, VGGNet, and ResNet have shown efficacy in machine learning competitions, research, and clinical use. Open-source machine learning libraries such as TensorFlow, pyTorch, and Keras are available for public use, fostering widespread adoption of this technology.
Figure 2. Diagram showing model development and application in the classification of medical images. The top row depicts the training process (A) and the bottom row the prediction process (B).
Figure 2. Diagram showing model development and application in the classification of medical images. The top row depicts the training process (A) and the bottom row the prediction process (B).
Bioengineering 11 00484 g002
Table 1. Selected articles; main characteristics.
Table 1. Selected articles; main characteristics.
AuthorsArtificial Intelligence MethodPublication YearMain ObjectivesTitle of JournalType of RadiographsPerformance
Ho, C.S. et al. [56]CNN2021Classify osteoporosisArch. Osteoporos.Pelvis and femurr = 0.850;
Accuracy 88.0%
Fathima, S.M.N. et al. [57]CNN (U-Net)2020Classify osteoporosisJ Xray Sci Technol.VariousAccuracy 88.0%;
Sensitivity 95.2–95.8%;
Specificity 96.7–97.5%
Hsieh, C.I. et al. [58]CNN (VGG-16 and ResNet-34)2021Classify osteoporosisNat. Commun.Lumbar spine, pelvisAUC 0.890;
Accuracy 86.2–91.7%;
Sensitivity 80.2%–83.5%;
Specificity 88.3%–94.9%
Sukegawa, S. et al. [59]CNN (EfficientNet-b0, -b3, and -b7 and ResNet-18, -50, and -152)2022Classify osteoporosisSci. Rep.Dental panoramicAUC 0.911–0.921;
Accuracy 84.0–84.5%;
Specificity 88.8–90.6%;
F1 score 0.720–0.740
Yamamoto, N. et al. [60]CNN (ResNet18, ResNet34, GoogleNet, EfficientNet b3, EfficientNetb4)2020Classify osteoporosisBiomoleculesHipAccuracy 88.5%;
Specificity 92.2%;
Recall 0.887;
F1 score 0.894;
AUC 0.922–0.937
Wani, I. et al. [30]CNN (AlexNet, VggNet-16, ResNet, VggNet-19)2022Classify osteoporosisMultimed. Tools Appl.KneeAccuracy 90.9%;
Error rate 9.0%;
Validation loss 54%
Lee, K.S. et al. [61]CNN (CNN3, VGG-16, VGG-16_TF, VGG-16_TF_FT)2020Predict osteoporosisJ Clin Med.Dental panoramicAUC 0.858;
Sensitivity 90.0%;
Specificity 81.5%;
Accuracy 84.0%
Zhang, B. et al. [62]CNN2020Classify osteoporosisBoneLumbar spineAUC 0.767–0.810;
Sensitivity 68.4–85.3%
Singh, A. et al. [32]SVM, GNB, k-NN, ANN2017Predict osteoporosisComput Biol Med.CalcaneumAUC 0.982;
Accuracy 97.9%;
Sensitivity 100.0%;
Specificity 95.7%
Tecle, N. et al. [31]CNN (FSN-8)2020Predict osteoporosisJ Hand Surg Am.HandSensitivity 82.4%;
Specificity 94.3%
Areeckal, A. S. et al. [63]k-NN2017Predict osteoporosisOsteoporos Int.Hand and wristAccuracy 93.2%;
Sensitivity 91.2%;
Specificity 95.0%
Kavitha, M.S. et al. [64]naive Bayes, k-NN, SVM2015Predict osteoporosisOral Surg Oral Med Oral Pathol Oral Radiol.Dental panoramicAccuracy 89.5–96.8%
Kavitha, M.S. et al. [65]hybrid GSF2016Classify osteoporosisDentomaxillofac Radiol.Dental panoramicAUC 0.986;
Sensitivity 99.1%;
Specificity 98.4%;
Accuracy 98.9% (femoral neck);
AUC 0.962; Sensitivity 95.3%;
Specificity 94.7%;
Accuracy 96.0% (lumbar spine)
Chu, P. et al. [66]CNN (OSN; AlexNet)2018Predict osteoporosisAnnu Int Conf IEEE Eng Med Biol SocDental panoramicAccuracy 89.8%
Hwang, J.J. et al. [67]decision tree, SVM2017Predict osteoporosisDentomaxillofac RadiolDental panoramicAccuracy 96.2–96.3%;
Sensitivity 97.1–97.2%;
Specificity 96.3–97.1
Lee, J.S. et al. [68]SC-DCNN, SC DNN Augment, MC-DCNN2019Predict osteoporosisDentomaxillofac RadiolDental panoramicAUC 0.973–0.999;
Accuracy 93.0–98.5%
Oulhaj. H. et al. [69]SVM2017Predict osteoporosisIEEE Trans Med ImagingCalcaneumAUC 0.930;
Accuracy 91.3%;
Sensitivity 92.0%;
Specificity 91.0%
Zheng, K. et al. [70]CNN (Alexnet, Googlenet, Resnet18, Inceptionv3)2020Predict osteoporosisArtif Intell MedCalcaneumAUC 0.944;
Accuracy 90.8%
Nasser, Y. et al. [71]SVM2017Predict osteoporosisNew York: IEEECalcaneumAccuracy 95.5%
Jang, M. et al. [34]Deep learning model (OsPor-screen)2022Classify osteoporosisJ. Bone Miner. Res.ChestAUC 0.880–0.910;
Accuracy 77.7–84.2%;
Sensitivity 84.3–86.2%;
Specificity 74.2–81.5%
Bhattacharya, S. et al. [72]SVM, NN2019Classify osteoporosisIEEE XploreCalcaneumAccuracy 95.6%
Jang, R. et al. [73]CNN (VGG16)2021Classify osteoporosisSci. Rep.HipAUC 0.700;
Accuracy 81.2%;
Sensitivity 91.1%;
Specificity 68.9%,
Nguyen, T. et al. [74]CNN2021Classify osteoporosisComput. Biol. Med.Hipr = 0.808
Singh, Y. et al. [75]CNN2021Classify osteoporosis43rd Conf Proc IEEE Eng Med Biol SocDental panoramicAccuracy: 87.9%
Sato, Y. et al. [35]CNN2022Classify osteoporosisBiomedicines.ChestAUC 0.700–0.890;
Accuracy 66.1–78.5%;
Sensitivity 71.3–90.1%;
Specificity 62.4–73.7%
Hong, N. et al. [76]CNN2023Classify osteoporosisJ Bone Miner Res.Lateral spineAUC 0.830–0.850;
Sensitivity 75.0–76.0%
Nakamoto, T. et al. [77]CNN2022Classify osteoporosisDentomaxillofac RadiolDental panoramicSensitivity 78.3–82.6%;
Specificity 71.4–79.2%;
Accuracy 74.0–79.0% (Lumbar spine DEXA);
Sensitivity 80.0–86.7%;
Specificity 67.1–74.1%;
Accuracy 70.0–75.0% (Femoral Neck)
Widyaningrum, R. et al. [51]DT, GNB, MLP2023Classify osteoporosisInt. J. Dent.Dental panaromicAccuracy 90.5%;
Specificity 90.9%;
Sensitivity 90.0%
Lee, S.W. et al. [78]CNN2020Classify osteoporosisSkeletal Radiol.SpineAUC 0.740;
Accuracy 71.0%;
Sensitivity 81.0%;
Specificity 60.0%;
F1-score 0.73
Mohammadi, F. G. et al. [79]CNN2023Classify osteoporosisStud Health Technol InformHandAUC 0.740;
Accuracy 82.0%;
Sensitivity 87.0%;
Specificity 61.0%
Mao, L. et al. [80]CNN2022Classify osteoporosisFront. Endocrinol.Lumbar spineAUC 0.937;
Sensitivity 84.8%;
Specificity 86.6%
Area under receiver operator curve (AUC), correlation coefficient (r).
Table 2. Results summary.
Table 2. Results summary.
Areas SampledNo. of StudiesAUCAccuracy (%)Sensitivity (%)Specificity (%)
Dental100.858–0.99974.0–96.978.3–97.267.1–97.1
Hip50.700–0.93781.2–88.580.2–91.168.9–94.9
Spine50.726–0.93771.0–86.268.4–84.860.0–88.3
Calcaneum50.930–0.98290.8–97.992.0–100.091.0–95.7
Hand or Wrist30.74082.0–93.282.4–91.261.0–95.7
Chest20.700–0.91066.1–84.271.3–90.162.4–81.5
Various *1-88.095.2–95.896.7–97.5
Knee1-90.9--
Overall320.700–0.99966.1–97.967.4–100.060.0–97.5
* Internal datasets were used consisting of spine, femur, knee, clavicle, and upper extremity radiographs.

4. Discussion

4.1. Advantages and Efficacy

Based on our comprehensive review, many machine-learning tools demonstrate impressive diagnostic capabilities for osteoporosis when benchmarked against established reference standards. Furthermore, these tools consistently show excellent discriminatory performance across various anatomical regions, with promising outcomes in osteoporosis prediction using radiographs of the hip, spine, chest, extremities, and mandible.
One major advantage is manpower and time savings for image segmentation and analysis. Segmentation is the process of identifying regions of interest (ROI) in images, such as separating bone and non-bone structures on radiographs. Historically, this was a manual and time-consuming task requiring trained personnel. In contrast, AI can handle vast imaging datasets without manual intervention, reducing human error and interobserver variability. As a case in point, Jang et al. noted that their automated “OsPor-screen” model required less than 4 s to process and classify a chest radiograph [34]. Similarly, Doctorant et al. described an AI model for ROI labeling in lumbar spine DEXA, which required only seconds for analysis and matched the performance of expert operators [52].
In addition, radiographs are an ideal modality for large-scale population screening due to their cost-effectiveness and typically lower radiation exposure when compared to DEXA and quantitative CT scans. Existing radiographs performed for other purposes may also be retrospectively analyzed for osteoporosis without incurring further costs or radiation burden to the patient, increasing the appeal of screening and improving screening program uptake in the population. Serial radiographs may also be performed for a patient over time depending on the clinical scenario, such as follow-up of chest infections and assessing fracture healing. This would permit the close trending of BMD without the need for frequent DEXA scans, although the clinical utility of this process is currently uncertain and requires further evaluation.
Finally, radiographs represent a promising alternative avenue for osteoporosis diagnosis in rural settings or developing countries where DEXA machines are not widely available, a role that is analogous to that of quantitative ultrasounds [81,82]. Given the high diagnostic accuracy of some deep learning models (AUCs up to 0.9987 in dental [68] and 0.9824 in calcaneal radiographs [32]), further research and validation is merited to establish if deep learning diagnosis of osteoporosis from radiographs in tandem with other clinical tools may suffice for treatment initiation in these underserved populations.

4.2. Challenges: Dataset Collection

The efficacy of a machine learning model is contingent on the size and quality of the dataset on which it is trained [48,49]. There are several challenges in medical image collection (Figure 3); first, data collection is subject to ethical considerations such as patient privacy and radiation exposure. Second, medical images can be very large in size with high resolutions. Finally, imaging platforms may not be readily amenable to data transfer and collection in view of security concerns [83]. Therefore, medical imaging datasets tend to be relatively small, comprising hundreds to thousands of images [48] compared to natural image datasets, which can contain millions of images.
Deep learning models trained on small datasets tend to have reduced generalizability and overfitting. To mitigate this problem, data augmentation techniques (such as blurring, shearing, sharpening, and rotation) are often performed to artificially increase the size of datasets without the need for further collection [84]. Transfer learning is also a relevant technique, whereby a CNN trained on a large dataset may be retrained on a smaller dataset for a new problem [30,85]. For example, Wani et al. [30] and Lee et al. [61] independently utilized transfer learning in the analysis of knee and dental radiographs, respectively, noting that the use of a pre-trained model helps to mitigate the problem of smaller datasets.
Dataset bias occurs when a dataset used to train a model has a different distribution from the population to which it is to be applied [86,87]. Some studies only examine a small subgroup of the population and thus may appear accurate if the test sets are derived from the same population but fail when tested in a broader context. A recent review of studies analyzing dental radiographs by Martins et al. [88] noted that most papers only incorporated data from a single institute, while Alberquerque et al. [89] noted the problem of dataset imbalance, whereby a dataset retrospectively including patients who had previously undergone DEXA scans would be skewed toward higher rates of osteoporosis than the general population. It can be difficult to determine if a dataset is subject to bias, particularly if the collection criteria are not disclosed. Critical evaluation of datasets by researchers and meticulous documentation of a dataset’s characteristics and patient demographics would help to alleviate this problem [83].

4.3. Challenges: Radiograph Quality and Confounding Pathologies

Another significant challenge lies in the technical quality of radiographs, as suboptimal positioning, variable imaging techniques, variations in image exposure, and the presence of artifacts can lead to inaccuracies in the model. One study by Socha et al. [90] examining COVID-19 detection AI models described how poor image quality, artifacts, and data heterogeneity in the initial datasets collected during the pandemic contributed to poor performance in real-world clinical settings. In the context of osteoporosis, Hsieh et al. [58] noted that bony pathologies such as fractures, implants, bony tumors, infections, and severe osteoarthritis can introduce complexity in analysis. It was postulated that fractures can alter normal bone anatomy and induce callus formation, whereas implants might produce metallic artifacts, rendering the evaluation of adjacent tissue more challenging. The presence of these factors necessitates algorithms and image processing methods to differentiate true BMD alterations from other non-BMD changes. Fortunately, the ubiquity of radiographs may partially circumvent this problem; for example, in a large AI pipeline, a patient with a hip implant may have radiographs of other anatomical regions such as the chest and other limbs performed at the same sitting and thereby undergo osteoporosis screening even though the hip radiograph may not be amenable to analysis [58].
The advent of AI tools for the detection of other bone pathologies may also help to mitigate the problem. For example, Hsieh et al. [58] incorporated multiple deep learning models in their automated osteoporosis detection pipeline, such as for the detection of fractures and other algorithms for the exclusion of poor-quality images and concomitant pathologies [58]. The tool was able to automatically exclude confounding radiographs and successfully report a predicted BMD in 79.0% of pelvis radiographs and 82.3% of spine radiographs in a large population of tertiary care patients at a general hospital without manual intervention.

4.4. Challenges: Study Protocol Heterogeneity

The variability in study protocols presents significant challenges in the field of AI-driven osteoporosis classification. Differences in the choice of inclusion/exclusion criteria, pre-processing methodology, and model construction make it difficult to compare findings across studies.
Moreover, inconsistencies in defining model endpoints and the choice of performance metrics further complicate the interpretation of results. For example, some papers opt to classify osteoporosis versus non-osteoporosis in patients while others endeavor to measure BMD directly. This diversity underscores the complexity of the issue and highlights the importance of standardizing methodologies to ensure consistency and reliability in research outcomes. Improved consistency in study protocols and standardizing methodologies will facilitate the accumulation of robust evidence, ultimately advancing our understanding of AI-driven solutions.

4.5. Challenges: Clinical Integration

The clinical integration of artificial intelligence solutions is fraught with challenges [91]. Recht et al. [92] outlined various ethical, technical and clinical challenges involved in the clinical integration of AI: AI algorithms must align with the complex and diverse spectrum of clinical protocols and practices across different regions and healthcare contexts. Data privacy and security concerns, along with the ethical implications of AI-driven decision support, also add layers of complexity to the integration process. Daye et al. [93] proposed a roadmap for successful oversight of clinical AI implementation, noting that four components are required for successful implementation: data access and security, cross-platform and cross-domain integration, clinical translation and delivery, and leadership supporting innovation.
Furthermore, the regulatory landscape is continuously evolving, necessitating frequent updates and adaptations to ensure AI systems comply with rigorous healthcare safety and quality standards [94]. Close cooperation with regulatory bodies such as the Food and Drug Administration in the United States is required [95].
Variability in data formats, quality, and acquisition techniques across various healthcare systems can also hinder the performance of AI tools. The heterogeneous distribution of disease in various populations and different populations also further complicates matters and may necessitate the use of separate training sets in different populations [90]. Most of the reviewed studies demonstrate good diagnostic accuracy on internal or external datasets without the inclusion of an integrated clinical pathway. On the other hand, Hsieh et al. [58]. outlined a process at the Chang Gung Memorial Hospital (Linkou, Taiwan) in which the hospital PACS relayed all newly acquired pelvic and lumbar spine radiographs to an inference platform daily [58]. The system also integrated several other deep learning tools for the detection of image quality and other bony pathologies such as hip or lumbar spine fractures and automatically excluded these studies. There is a clear need for further research and infrastructure development to support the establishment of similar integrated clinical platforms.
As a counterpoint, various studies do make use of clinical data to improve diagnostic accuracy. For example, Yamato et al. [60], Sukegawa et al. [59], and Mao et al. [80] describe the use of ensemble models incorporating clinical covariate data such as age, gender, and BMI. Such ensemble models generally improved model performance on various metrics, especially accuracy and AUC [59,60].
Proving the effectiveness of a deep learning model is also often challenging since it often functions like a black box. However, recent visualization methods, such as Grad-CAM [59,96] and back-propagation visualization [97], highlight salient areas of interest in images and may help to increase the trust and acceptance of both patients and clinicians in the AI model. For example, Jang et al. [34] described the use of Grad-CAM in their paper on chest radiographs and were able to present graphical illustrations of how the model made positive predictions based on various locations such as the humeral head, scapula, ribs, spine, and clavicle; it was noted that further work on visualization methods may help to improve the interpretability of AI models and improve clinician acceptance.

4.6. Future Directions in the Use of AI in Osteoporosis

There are numerous exciting opportunities for harnessing AI to predict osteoporosis from medical imaging. Osteoporosis may serve as an ideal stepping stone for the introduction of automated imaging systems to healthcare because the disease is highly prevalent with well-established benchmarks and reference standards in the form of DEXA but also not time-critical and amenable to further verification and follow-up. These automated platforms can then be integrated with other deep learning tools such as the detection of image quality, fractures, implants, and bony tumors that can help to reduce confounding factors for osteoporosis detection as well as serve as clinically relevant diagnostic tools in their own right.
Multimodality platforms may also serve as a useful direction for exploration. The use of CT assessment of osteoporosis is well established in the literature, with various tools showing excellent diagnostic performance [98,99]. MRI tools may also be useful: Zhao et al. proposed a fully automated radiomic screening pipeline for osteoporosis using a short lumbar mDIXON sequence for opportunistic screening, which could be performed in as short as 16 s [100]. A combined platform harnessing multiple modalities could yield higher accuracy with lower patient costs.
There is also great interest in the use of deep learning techniques to uncover further risk factors and predict fracture risk independent of BMD and FRAX [101]. For example, Yosibash et al. [102] described an autonomous algorithm combining autonomous finite element analysis and machine learning techniques for accurate prediction of future hip fracture risk assessment from CT scans of the abdomen and pelvis [102]. Further investigation and exploration of clinical integration of these methods is warranted.
Radiologists and clinicians must actively engage in the training and adoption of AI technologies and healthcare institutions must invest in infrastructure and education to support this transformative shift in medical imaging. Ultimately, a collaborative effort between clinicians, technology developers, and regulatory bodies is crucial to overcoming these challenges and realizing the full potential of AI in radiology.

5. Conclusions

This systematic review highlights the growing body of evidence that underscores the promise of harnessing artificial intelligence in radiographs for osteoporosis classification. Modern deep-learning technology allows for the automated analysis of substantial volumes of radiographic data, eliminating the need for labor-intensive manual segmentation and image analysis. In addition, the cost-effectiveness and accessibility of radiographs make them an ideal modality for large-scale population screening, particularly in settings where DEXA machines may be scarce. By leveraging the high diagnostic accuracy of deep learning models, there is a promising opportunity to enhance osteoporosis diagnosis and treatment initiation in these populations, thereby improving healthcare equity.
Several challenges must be addressed to fully realize the potential of AI-driven osteoporosis classification. Dataset collection is a major hurdle due to ethical concerns, data privacy issues, and biases. Standardizing methodologies and rigorously evaluating datasets are crucial for reliable and generalizable AI models across diverse populations. Moreover, ensuring robust algorithms capable of accurately distinguishing true bone mineral density alterations from other changes is essential given the technical quality of radiographs and the presence of confounding pathologies.
Looking ahead, collaboration between radiologists, clinicians, technology developers, and regulatory bodies is crucial to overcome the challenges associated with AI implementation and ensure patient-centric care. Future research endeavors should focus on addressing the challenges in technical application and clinical integration to facilitate future practical implementation of this technology.

Author Contributions

Writing, supervision, methodology, and conceptualization: J.T.P.D.H., R.W.L., W.O., J.H.T., A.M. and N.K.; Project administration and investigation: R.W.L., W.O. and J.T.P.D.H.; Resources and software: R.W.L., W.O., A.M., X.Z.L., G.S., T.Y.L., D.F.K.T. and J.T.P.D.H.; Formal analysis and validation: J.T.P.D.H., R.W.L. and W.O. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by direct funding from MOH/NMRC. This research is supported by the Singapore Ministry of Health National Medical Research Council under the NMRC Clinician Innovator Award (CIA); Grant Title: Deep learning pipeline for augmented reporting of MRI whole spine (CIAINV23jan-0001, MOH-001405).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Föger-Samwald, U.; Dovjak, P.; Azizi-Semrad, U.; Kerschan-Schindl, K.; Pietschmann, P. Osteoporosis: Pathophysiology and therapeutic options. EXCLI J. 2020, 19, 1017–1037. [Google Scholar] [CrossRef] [PubMed]
  2. Sheik Ali, A. Osteoporosis: A Narrative Review. Cureus 2023, 15, e43031. [Google Scholar] [CrossRef] [PubMed]
  3. Curry, S.J.; Krist, A.H.; Owens, D.K.; Barry, M.J.; Caughey, A.B.; Davidson, K.W.; Doubeni, C.A.; Epling, J.W.; Kemper, A.R.; Kubik, M.; et al. Screening for Osteoporosis to Prevent Fractures. JAMA 2018, 319, 2521. [Google Scholar] [CrossRef] [PubMed]
  4. Brooke-Wavell, K.; Skelton, D.A.; Barker, K.L.; Clark, E.M.; De Biase, S.; Arnold, S.; Paskins, Z.; Robinson, K.R.; Lewis, R.M.; Tobias, J.H.; et al. Strong, steady and straight: UK consensus statement on physical activity and exercise for osteoporosis. Br. J. Sports Med. 2022, 56, 837–846. [Google Scholar] [CrossRef] [PubMed]
  5. Keen, M.U.; Reddivari, A.K.R. Osteoporosis in Females; StatPearls: St. Petersburg, FL, USA, 2024. [Google Scholar]
  6. Björnsdottir, S.; Clarke, B.L.; Mannstadt, M.; Langdahl, B.L. Male osteoporosis-what are the causes, diagnostic challenges, and management. Best Pract. Res. Clin. Rheumatol. 2022, 36, 101766. [Google Scholar] [CrossRef] [PubMed]
  7. Kanis, J.A.; Johnell, O.; Oden, A.; Sernbo, I.; Redlund-Johnell, I.; Dawson, A.; De Laet, C.; Jonsson, B. Long-Term Risk of Osteoporotic Fracture in Malmö. Osteoporos. Int. 2000, 11, 669–674. [Google Scholar] [CrossRef] [PubMed]
  8. Kanis, J. Assessment of osteoporosis at the primary health-care level. In WHO Scientific Group Technical Report. 2007; WHO Collaborating Centre for Metabolic Bone Diseases, University of Sheffield: Sheffield, UK, 2007. [Google Scholar]
  9. Leibson, C.L.; Tosteson, A.N.A.; Gabriel, S.E.; Ransom, J.E.; Melton, L.J. Mortality, Disability, and Nursing Home Use for Persons with and without Hip Fracture: A Population-Based Study. J. Am. Geriatr. Soc. 2002, 50, 1644–1650. [Google Scholar] [CrossRef] [PubMed]
  10. Clynes, M.A.; Harvey, N.C.; Curtis, E.M.; Fuggle, N.R.; Dennison, E.M.; Cooper, C. The epidemiology of osteoporosis. Br. Med. Bull. 2020, 133, 105–117. [Google Scholar] [CrossRef] [PubMed]
  11. Wáng, Y.X.J.; Lentle, B.C. Radiographic osteoporotic vertebral fractures in elderly men: A brief review focusing on differences between the sexes. Quant. Imaging Med. Surg. 2020, 10, 1863–1876. [Google Scholar] [CrossRef] [PubMed]
  12. Office of the Surgeon General (US). Bone Health and Osteoporosis: A Report of the Surgeon General; Office of the Surgeon General (US): Rockville, MD, USA, 2004. [Google Scholar]
  13. Burge, R.; Dawson-Hughes, B.; Solomon, D.H.; Wong, J.B.; King, A.; Tosteson, A. Incidence and Economic Burden of Osteoporosis-Related Fractures in the United States, 2005–2025. J. Bone Miner. Res. 2007, 22, 465–475. [Google Scholar] [CrossRef]
  14. Hernlund, E.; Svedbom, A.; Ivergård, M.; Compston, J.; Cooper, C.; Stenmark, J.; McCloskey, E.V.; Jonsson, B.; Kanis, J.A. Osteoporosis in the European Union: Medical management, epidemiology and economic burden. Arch. Osteoporos. 2013, 8, 136. [Google Scholar] [CrossRef] [PubMed]
  15. Barr, R.J.; Stewart, A.; Torgerson, D.J.; Reid, D.M. Population screening for osteoporosis risk: A randomised control trial of medication use and fracture risk. Osteoporos. Int. 2010, 21, 561–568. [Google Scholar] [CrossRef] [PubMed]
  16. Miller, P.D. Underdiagnoses and Undertreatment of Osteoporosis: The Battle to Be Won. J. Clin. Endocrinol. Metab. 2016, 101, 852–859. [Google Scholar] [CrossRef] [PubMed]
  17. Lulla, D.; Teo, C.; Shen, X.; Loi, Z.; Quek, K.; Lis, H.; Koh, S.; Chan, E.; Lim, S.; Low, L. Assessing the knowledge, attitude and practice of osteoporosis among Singaporean women aged 65 years and above at two SingHealth polyclinics. Singap. Med. J. 2021, 62, 190–194. [Google Scholar] [CrossRef] [PubMed]
  18. Singer, A.; McClung, M.R.; Tran, O.; Morrow, C.D.; Goldstein, S.; Kagan, R.; McDermott, M.; Yehoshua, A. Treatment rates and healthcare costs of patients with fragility fracture by site of care: A real-world data analysis. Arch. Osteoporos. 2023, 18, 42. [Google Scholar] [CrossRef] [PubMed]
  19. Cummings, S.R.; Bates, D.; Black, D.M. Clinical Use of Bone Densitometry. JAMA 2002, 288, 1889. [Google Scholar] [CrossRef] [PubMed]
  20. Lewiecki, E.M.; Compston, J.E.; Miller, P.D.; Adachi, J.D.; Adams, J.E.; Leslie, W.D.; Kanis, J.A.; Moayyeri, A.; Adler, R.A.; Hans, D.B.; et al. Official Positions for FRAX® Bone Mineral Density and FRAX® Simplification. J. Clin. Densitom. 2011, 14, 226–236. [Google Scholar] [CrossRef] [PubMed]
  21. Kanis, J.A.; Johnell, O. Requirements for DXA for the management of osteoporosis in Europe. Osteoporos. Int. 2005, 16, 229–238. [Google Scholar] [CrossRef] [PubMed]
  22. Engelke, K. Quantitative Computed Tomography—Current Status and New Developments. J. Clin. Densitom. 2017, 20, 309–321. [Google Scholar] [CrossRef] [PubMed]
  23. Schousboe, J.T.; Riekkinen, O.; Karjalainen, J. Prediction of hip osteoporosis by DXA using a novel pulse-echo ultrasound device. Osteoporos. Int. 2017, 28, 85–93. [Google Scholar] [CrossRef] [PubMed]
  24. Sung, K.H.; Choi, Y.; Cho, G.H.; Chung, C.Y.; Park, M.S.; Lee, K.M. Peripheral DXA measurement around ankle joint to diagnose osteoporosis as assessed by central DXA measurement. Skelet. Radiol. 2018, 47, 1111–1117. [Google Scholar] [CrossRef] [PubMed]
  25. Barnett, E.; Nordin, B.E.C. The radiological diagnosis of osteoporosis: A new approach. Clin. Radiol. 1960, 11, 166–174. [Google Scholar] [CrossRef] [PubMed]
  26. Exton-Smith, A.N.; Millard, P.H.; Payne Erica, P.R.; Wheeler, F. Method for Measuring Quantity of Bone. Lancet 1969, 294, 1153–1154. [Google Scholar] [CrossRef] [PubMed]
  27. Garn, S.M. An Annotated Bibliography on Bone Densitometry. Am. J. Clin. Nutr. 1962, 10, 59–67. [Google Scholar] [CrossRef] [PubMed]
  28. Engstrom, A.; Welin, S. A method for the quantitative roentgenological determination of the amount of calcium salts in bone tissue. Acta Radiol. 1949, 31, 483–502. [Google Scholar] [CrossRef] [PubMed]
  29. Bywaters, E.G.L. The measurement of bone opacity. Clin. Sci. 1948, 6, 281–287. [Google Scholar] [PubMed]
  30. Wani, I.M.; Arora, S. Osteoporosis diagnosis in knee X-rays by transfer learning based on convolution neural network. Multimed. Tools Appl. 2023, 82, 14193–14217. [Google Scholar] [CrossRef] [PubMed]
  31. Tecle, N.; Teitel, J.; Morris, M.R.; Sani, N.; Mitten, D.; Hammert, W.C. Convolutional Neural Network for Second Metacarpal Radiographic Osteoporosis Screening. J. Hand Surg. 2020, 45, 175–181. [Google Scholar] [CrossRef] [PubMed]
  32. Singh, A.; Dutta, M.K.; Jennane, R.; Lespessailles, E. Classification of the trabecular bone structure of osteoporotic patients using machine vision. Comput. Biol. Med. 2017, 91, 148–158. [Google Scholar] [CrossRef] [PubMed]
  33. Cui, J.; Liu, C.L.; Jennane, R.; Ai, S.; Dai, K.; Tsai, T.-Y. A highly generalized classifier for osteoporosis radiography based on multiscale fractal, lacunarity, and entropy distributions. Front. Bioeng. Biotechnol. 2023, 11, 1054991. [Google Scholar] [CrossRef] [PubMed]
  34. Jang, M.; Kim, M.; Bae, S.J.; Lee, S.H.; Koh, J.-M.; Kim, N. Opportunistic Osteoporosis Screening Using Chest Radiographs with Deep Learning: Development and External Validation with a Cohort Dataset. J. Bone Miner. Res. 2022, 37, 369–377. [Google Scholar] [CrossRef] [PubMed]
  35. Sato, Y.; Yamamoto, N.; Inagaki, N.; Iesaki, Y.; Asamoto, T.; Suzuki, T. Deep Learning for Bone Mineral Density and T-Score Prediction from Chest X-rays: A Multicenter Study. Biomedicines 2022, 10, 2323. [Google Scholar] [CrossRef] [PubMed]
  36. Liu, X.; Faes, L.; Kale, A.U.; Wagner, S.K.; Fu, D.J.; Bruynseels, A.; Mahendiran, T.; Moraes, G.; Shamdas, M.; Kern, C.; et al. A comparison of deep learning performance against health-care professionals in detecting diseases from medical imaging: A systematic review and meta-analysis. Lancet Digit. Health 2019, 1, e271–e297. [Google Scholar] [CrossRef] [PubMed]
  37. Sendak, M.P.; D’Arcy, J.; Kashyap, S.; Gao, M.; Nichols, M.; Corey, K.; Ratliff, W.; Balu, S. A Path for Translation of Machine Learning Products into Healthcare Delivery. EMJ Innov. 2020, 10, 19–00172. [Google Scholar] [CrossRef]
  38. Topol, E.J. High-performance medicine: The convergence of human and artificial intelligence. Nat. Med. 2019, 25, 44–56. [Google Scholar] [CrossRef] [PubMed]
  39. Lind Plesner, L.; Müller, F.C.; Brejnebøl, M.W.; Laustrup, L.C.; Rasmussen, F.; Nielsen, O.W.; Boesen, M.; Andersen, M.B. Commercially Available Chest Radiograph AI Tools for Detecting Airspace Disease, Pneumothorax, and Pleural Effusion. Radiology 2023, 308, e231236. [Google Scholar] [CrossRef] [PubMed]
  40. Dembrower, K.; Crippa, A.; Colón, E.; Eklund, M.; Strand, F.; ScreenTrustCAD Trial Consortium. Artificial intelligence for breast cancer detection in screening mammography in Sweden: A prospective, population-based, paired-reader, non-inferiority study. Lancet. Digit. Health 2023, 5, e703–e711. [Google Scholar] [CrossRef] [PubMed]
  41. Hallinan, J.T.P.D.; Zhu, L.; Yang, K.; Makmur, A.; Algazwi, D.A.R.; Thian, Y.L.; Lau, S.; Choo, Y.S.; Eide, S.E.; Yap, Q.V.; et al. Deep Learning Model for Automated Detection and Classification of Central Canal, Lateral Recess, and Neural Foraminal Stenosis at Lumbar Spine MRI. Radiology 2021, 300, 130–138. [Google Scholar] [CrossRef] [PubMed]
  42. Lim, D.S.W.; Makmur, A.; Zhu, L.; Zhang, W.; Cheng, A.J.L.; Sia, D.S.Y.; Eide, S.E.; Ong, H.Y.; Jagmohan, P.; Tan, W.C.; et al. Improved Productivity Using Deep Learning-assisted Reporting for Lumbar Spine MRI. Radiology 2022, 305, 160–166. [Google Scholar] [CrossRef] [PubMed]
  43. Dercle, L.; McGale, J.; Sun, S.; Marabelle, A.; Yeh, R.; Deutsch, E.; Mokrane, F.-Z.; Farwell, M.; Ammari, S.; Schoder, H.; et al. Artificial intelligence and radiomics: Fundamentals, applications, and challenges in immunotherapy. J. ImmunoTher. Cancer 2022, 10, e005292. [Google Scholar] [CrossRef]
  44. Kocak, B.; Durmaz, E.S.; Ates, E.; Kilickesmez, O. Radiomics with artificial intelligence: A practical guide for beginners. Diagn. Interv. Radiol. 2019, 25, 485–495. [Google Scholar] [CrossRef] [PubMed]
  45. Erickson, B.J.; Korfiatis, P.; Akkus, Z.; Kline, T.L. Machine Learning for Medical Imaging. RadioGraphics 2017, 37, 505–515. [Google Scholar] [CrossRef] [PubMed]
  46. Sarker, I.H. Machine Learning: Algorithms, Real-World Applications and Research Directions. SN Comput. Sci. 2021, 2, 160. [Google Scholar] [CrossRef] [PubMed]
  47. Montagnon, E.; Cerny, M.; Cadrin-Chênevert, A.; Hamilton, V.; Derennes, T.; Ilinca, A.; Vandenbroucke-Menu, F.; Turcotte, S.; Kadoury, S.; Tang, A. Deep learning workflow in radiology: A primer. Insights Into Imaging 2020, 11, 22. [Google Scholar] [CrossRef] [PubMed]
  48. Willemink, M.J.; Koszek, W.A.; Hardell, C.; Wu, J.; Fleischmann, D.; Harvey, H.; Folio, L.R.; Summers, R.M.; Rubin, D.L.; Lungren, M.P. Preparing Medical Imaging Data for Machine Learning. Radiology 2020, 295, 4–15. [Google Scholar] [CrossRef] [PubMed]
  49. Erickson, B.J.; Fajnwaks, P.; Langer, S.G.; Perry, J. Multisite Image Data Collection and Management Using the RSNA Image Sharing Network. Transl. Oncol. 2014, 7, 36–39. [Google Scholar] [CrossRef] [PubMed]
  50. Hesamian, M.H.; Jia, W.; He, X.; Kennedy, P. Deep Learning Techniques for Medical Image Segmentation: Achievements and Challenges. J. Digit. Imaging 2019, 32, 582–596. [Google Scholar] [CrossRef] [PubMed]
  51. Widyaningrum, R.; Sela, E.I.; Pulungan, R.; Septiarini, A. Automatic Segmentation of Periapical Radiograph Using Color Histogram and Machine Learning for Osteoporosis Detection. Int. J. Dent. 2023, 2023, 6662911. [Google Scholar] [CrossRef] [PubMed]
  52. Doctorant, G.G.; Lebrahim, E.H.A.; Hind, K.; Olivier, L.; Rodriguez, E.G.; Beaugé, L.; Hans, D. Evaluation of AI-based spine segmentation (SpS) for the analysis of lumbar spine dual energy X-ray absorptiometry scans. J. Clin. Densitom. 2023, 26, 101410. [Google Scholar] [CrossRef]
  53. Kocak, B. Key concepts, common pitfalls, and best practices in artificial intelligence and machine learning: Focus on radiomics. Diagn. Interv. Radiol. 2022, 28, 450–462. [Google Scholar] [CrossRef] [PubMed]
  54. Sarvamangala, D.R.; Kulkarni, R.V. Convolutional neural networks in medical image understanding: A survey. Evol. Intell. 2022, 15, 1–22. [Google Scholar] [CrossRef] [PubMed]
  55. Yadav, S.S.; Jadhav, S.M. Deep convolutional neural network based medical image classification for disease diagnosis. J. Big Data 2019, 6, 113. [Google Scholar] [CrossRef]
  56. Ho, C.S.; Chen, Y.P.; Fan, T.Y.; Kuo, C.F.; Yen, T.Y.; Liu, Y.C.; Pei, Y.C. Application of deep learning neural network in predicting bone mineral density from plain X-ray radiography. Arch Osteoporos 2021, 16, 1–12. [Google Scholar] [CrossRef] [PubMed]
  57. Nazia Fathima, S.M.; Tamilselvi, R.; Parisa Beham, M.; Sabarinathan, D. Diagnosis of Osteoporosis using modified U-net architecture with attention unit in DEXA and X-ray images. J. X-ray Sci. Technol. 2020, 28, 953–973. [Google Scholar] [CrossRef] [PubMed]
  58. Hsieh, C.I.; Zheng, K.; Lin, C.; Mei, L.; Lu, L.; Li, W. Automated bone mineral density prediction and fracture risk assessment using plain radiographs via deep learning. Nat. Commun. 2021, 12, 5472. [Google Scholar] [CrossRef]
  59. Sukegawa, S.; Fujimura, A.; Taguchi, A.; Yamamoto, N.; Kitamura, A.; Goto, R. Identification of osteoporosis using ensemble deep learning model with panoramic radiographs and clinical covariates. Sci. Rep. 2022, 12, 6088. [Google Scholar] [CrossRef] [PubMed]
  60. Yamamoto, N.; Sukegawa, S.; Kitamura, A.; Goto, R.; Noda, T.; Nakano, K. Deep Learning for Osteoporosis Classification Using Hip Radiographs and Patient Clinical Covariates. Biomolecules 2020, 10, 1534. [Google Scholar] [CrossRef] [PubMed]
  61. Lee, K.S.; Jung, S.K.; Ryu, J.J.; Shin, S.W.; Choi, J. Evaluation of Transfer Learning with Deep Convolutional Neural Networks for Screening Osteoporosis in Dental Panoramic Radiographs. J. Clin. Med. 2020, 9, 392. [Google Scholar] [CrossRef] [PubMed]
  62. Zhang, B.; Yu, K.; Ning, Z.; Wang, K.; Dong, Y.; Liu, X. Deep learning of lumbar spine X-ray for osteopenia and osteoporosis screening: A multicenter retrospective cohort study. Bone 2020, 140, 115561. [Google Scholar] [CrossRef] [PubMed]
  63. Areeckal, A.S.; Jayasheelan, N.; Kamath, J.; Zawadynski, S.; Kocher, M.; David, S.S. Early diagnosis of osteoporosis using radiogrammetry and texture analysis from hand and wrist radiographs in Indian population. Osteoporos. Int. 2018, 29, 665–673. [Google Scholar] [CrossRef]
  64. Kavitha, M.S.; An, S.Y.; An, C.H.; Huh, K.H.; Yi, W.J.; Heo, M.S. Texture analysis of mandibular cortical bone on digital dental panoramic radiographs for the diagnosis of osteoporosis in Korean women. Oral Surg. Oral Med. Oral Pathol. Oral Radiol. 2015, 119, 346–356. [Google Scholar] [CrossRef] [PubMed]
  65. Kavitha, M.S.; Ganesh Kumar, P.; Park, S.Y.; Huh, K.H.; Heo, M.S.; Kurita, T. Automatic detection of osteoporosis based on hybrid genetic swarm fuzzy classifier approaches. Dentomaxillofac. Radiol. 2016, 45, 20160076. [Google Scholar] [CrossRef] [PubMed]
  66. Chu, P.; Bo, C.; Liang, X.; Yang, J.; Megalooikonomou, V.; Yang, F. Using Octuplet Siamese Network For Osteoporosis Analysis On Dental Panoramic Radiographs. In Proceedings of the 2018 40th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Honolulu, HI, USA, 18–21 July 2018; Volume 2018, pp. 2579–2582. [Google Scholar]
  67. Hwang, J.J.; Lee, J.H.; Han, S.S.; Kim, Y.H.; Jeong, H.G.; Choi, Y.J. Strut analysis for osteoporosis detection model using dental panoramic radiography. Dentomaxillofac. Radiol. 2017, 46, 20170006. [Google Scholar] [CrossRef] [PubMed]
  68. Lee, J.S.; Adhikari, S.; Liu, L.; Jeong, H.G.; Kim, H.; Yoon, S.J. Osteoporosis detection in panoramic radiographs using a deep convolutional neural network-based computer-assisted diagnosis system: A preliminary study. Dentomaxillofac. Radiol. 2019, 48, 20170344. [Google Scholar] [CrossRef] [PubMed]
  69. Oulhaj, H.; Rziza, M.; Amine, A.; Toumi, H.; Lespessailles, E.; Hassouni, M.E. Anisotropic Discrete Dual-Tree Wavelet Transform for Improved Classification of Trabecular Bone. IEEE Trans. Med. Imaging 2017, 36, 2077–2086. [Google Scholar] [CrossRef] [PubMed]
  70. Zheng, K.; Harris, C.E.; Jennane, R.; Makrogiannis, S. Integrative blockwise sparse analysis for tissue characterization and classification. Artif. Intell. Med. 2020, 107, 101885. [Google Scholar] [CrossRef] [PubMed]
  71. Nasser, Y.; Hassouni, M.E.; Brahim, A.; Toumi, H.; Lespessailles, E. Diagnosis of osteoporosis disease from bone X-ray images with stacked sparse autoencoder and SVM classifier. In 2017 International Conference on Advanced Technologies for Signal and Image Processing (ATSIP), Fez, Morocco, 22–24 May 2017; Jennane, R., Ed.; IEEE Xplore: Piscataway, NJ, USA, 2017. [Google Scholar]
  72. Bhattacharya, S.; Nair, D.; Bhan, A.; Goyal, A. Computer Based Automatic Detection and Classification of Osteoporosis in Bone Radiographs. In Proceedings of the 2019 6th International Conference on Signal Processing and Integrated Networks (SPIN), Noida, India, 7–8 March 2019; IEEE: New York, NY, USA, 2019; pp. 1047–1052. [Google Scholar] [CrossRef]
  73. Jang, R.; Choi, J.H.; Kim, N.; Chang, J.S.; Yoon, P.W.; Kim, C.-H. Prediction of osteoporosis from simple hip radiography using deep learning algorithm. Sci. Rep. 2021, 11, 19997. [Google Scholar] [CrossRef] [PubMed]
  74. Nguyen, T.P.; Chae, D.-S.; Park, S.-J.; Yoon, J. A novel approach for evaluating bone mineral density of hips based on Sobel gradient-based map of radiographs utilizing convolutional neural network. Comput. Biol. Med. 2021, 132, 104298. [Google Scholar] [CrossRef] [PubMed]
  75. Singh, Y.; Atulkar, V.; Ren, J.; Yang, J.; Fan, H.; Latecki, L.J.; Ling, H. Osteoporosis Prescreening and Bone Mineral Density Prediction using Dental Panoramic Radiographs. In Proceedings of the 2021 43rd Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), Mexico City, Mexico, 1–5 November 2021; IEEE: New York, NY, USA, 2021; pp. 2700–2703. [Google Scholar] [CrossRef]
  76. Hong, N.; Cho, S.W.; Shin, S.; Lee, S.; Jang, S.A.; Roh, S. Deep-Learning-Based Detection of Vertebral Fracture and Osteoporosis Using Lateral Spine X-ray Radiography. J. Bone Miner. Res. 2023, 38, 887–895. [Google Scholar] [CrossRef] [PubMed]
  77. Nakamoto, T.; Taguchi, A.; Kakimoto, N. Osteoporosis screening support system from panoramic radiographs using deep learning by convolutional neural network. Dentomaxillofac. Radiol. 2022, 51, 20220135. [Google Scholar] [CrossRef] [PubMed]
  78. Lee, S.; Choe, E.K.; Kang, H.Y.; Yoon, J.W.; Kim, H.S. The exploration of feature extraction and machine learning for predicting bone density from simple spine X-ray images in a Korean population. Skelet. Radiol. 2020, 49, 613–618. [Google Scholar] [CrossRef] [PubMed]
  79. Mohammadi, F.G.; Sebro, R. Opportunistic Screening for Osteoporosis Using Hand Radiographs: A Preliminary Study. Stud. Health Technol. Inf. 2023, 302, 911–912. [Google Scholar]
  80. Mao, L.; Xia, Z.; Pan, L.; Chen, J.; Liu, X.; Li, Z. Deep learning for screening primary osteopenia and osteoporosis using spine radiographs and patient clinical covariates in a Chinese population. Front. Endocrinol. 2022, 13, 971877. [Google Scholar] [CrossRef] [PubMed]
  81. Nayak, S.; Olkin, I.; Liu, H.; Grabe, M.; Gould, M.K.; Allen, I.E.; Owens, D.K.; Bravata, D.M. Meta-Analysis: Accuracy of Quantitative Ultrasound for Identifying Patients with Osteoporosis. Ann. Intern. Med. 2006, 144, 832. [Google Scholar] [CrossRef] [PubMed]
  82. Moayyeri, A.; Adams, J.E.; Adler, R.A.; Krieg, M.-A.; Hans, D.; Compston, J.; Lewiecki, E.M. Quantitative ultrasound of the heel and fracture risk assessment: An updated meta-analysis. Osteoporos. Int. 2012, 23, 143–153. [Google Scholar] [CrossRef] [PubMed]
  83. Varoquaux, G.; Cheplygina, V. Machine learning for medical imaging: Methodological failures and recommendations for the future. NPJ Digit. Med. 2022, 5, 48. [Google Scholar] [CrossRef] [PubMed]
  84. Chlap, P.; Min, H.; Vandenberg, N.; Dowling, J.; Holloway, L.; Haworth, A. A review of medical image data augmentation techniques for deep learning applications. J. Med. Imaging Radiat. Oncol. 2021, 65, 545–563. [Google Scholar] [CrossRef] [PubMed]
  85. Torrey, L.; Shavlik, J. Transfer Learning. In Handbook of Research on Machine Learning Applications and Trends; IGI Global: Hershey, PA, USA, 2010; pp. 242–264. [Google Scholar] [CrossRef]
  86. Ricci Lara, M.A.; Echeveste, R.; Ferrante, E. Addressing fairness in artificial intelligence for medical imaging. Nat. Commun. 2022, 13, 4581. [Google Scholar] [CrossRef] [PubMed]
  87. Banerjee, I.; Bhattacharjee, K.; Burns, J.L.; Trivedi, H.; Purkayastha, S.; Seyyed-Kalantari, L.; Patel, B.N.; Shiradkar, R.; Gichoya, J. “Shortcuts” Causing Bias in Radiology Artificial Intelligence: Causes, Evaluation, and Mitigation. J. Am. Coll. Radiol. 2023, 20, 842–851. [Google Scholar] [CrossRef]
  88. Martins, M.V.; Baptista, L.; Luís, H.; Assunção, V.; Araújo, M.-R.; Realinho, V. Machine Learning in X-ray Diagnosis for Oral Health: A Review of Recent Progress. Computation 2023, 11, 115. [Google Scholar] [CrossRef]
  89. Albuquerque, G.A.; Carvalho, D.D.A.; Cruz, A.S.; Santos, J.P.Q.; Machado, G.M.; Gendriz, I.S.; Fernandes, F.R.S.; Barbalho, I.M.P.; Santos, M.M.; Teixeira, C.A.D.; et al. Osteoporosis screening using machine learning and electromagnetic waves. Sci. Rep. 2023, 13, 12865. [Google Scholar] [CrossRef] [PubMed]
  90. Socha, M.; Prażuch, W.; Suwalska, A.; Foszner, P.; Tobiasz, J.; Jaroszewicz, J.; Gruszczynska, K.; Sliwinska, M.; Nowak, M.; Gizycka, B.; et al. Pathological changes or technical artefacts? The problem of the heterogenous databases in COVID-19 CXR image analysis. Comput. Methods Programs Biomed. 2023, 240, 107684. [Google Scholar] [CrossRef] [PubMed]
  91. Kotter, E.; Ranschaert, E. Challenges and solutions for introducing artificial intelligence (AI) in daily clinical workflow. Eur. Radiol. 2021, 31, 5–7. [Google Scholar] [CrossRef] [PubMed]
  92. Recht, M.P.; Dewey, M.; Dreyer, K.; Langlotz, C.; Niessen, W.; Prainsack, B.; Smith, J.J. Integrating artificial intelligence into the clinical practice of radiology: Challenges and recommendations. Eur. Radiol. 2020, 30, 3576–3584. [Google Scholar] [CrossRef] [PubMed]
  93. Daye, D.; Wiggins, W.F.; Lungren, M.P.; Alkasab, T.; Kottler, N.; Allen, B.; Roth, C.J.; Bizzo, B.C.; Durniak, K.; Brink, J.A.; et al. Implementation of Clinical Artificial Intelligence in Radiology: Who Decides and How? Radiology 2022, 305, 555–563. [Google Scholar] [CrossRef] [PubMed]
  94. Pesapane, F.; Suter, M.B.; Codari, M.; Patella, F.; Volonté, C.; Sardanelli, F. Regulatory issues for artificial intelligence in radiology. In Precision Medicine for Investigators, Practitioners and Providers; Elsevier: Amsterdam, The Netherlands, 2020; pp. 533–543. [Google Scholar] [CrossRef]
  95. Petrick, N.; Chen, W.; Delfino, J.G.; Gallas, B.D.; Kang, Y.; Krainak, D.; Sahiner, B.; Samala, R.K. Regulatory considerations for medical imaging AI/ML devices in the United States: Concepts and challenges. J. Med. Imaging 2023, 10, 051804. [Google Scholar] [CrossRef] [PubMed]
  96. Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-CAM: Visual Explanations from Deep Networks via Gradient-Based Localization. Int. J. Comput. Vis. 2020, 128, 336–359. [Google Scholar] [CrossRef]
  97. Springenberg, J.T.; Dosovitskiy, A.; Brox, T.; Riedmiller, M. Striving for Simplicity: The All Convolutional Net. arxiv 2014, arXiv:1412.6806. [Google Scholar]
  98. Rahim, F.; Zaki Zadeh, A.; Javanmardi, P.; Emmanuel Komolafe, T.; Khalafi, M.; Arjomandi, A.; Ghofrani, H.A.; Shirbandi, K. Machine learning algorithms for diagnosis of hip bone osteoporosis: A systematic review and meta-analysis study. BioMed. Eng. OnLine 2023, 22, 68. [Google Scholar] [CrossRef] [PubMed]
  99. Pickhardt, P.J.; Nguyen, T.; Perez, A.A.; Graffy, P.M.; Jang, S.; Summers, R.M.; Garrett, J.W. Improved CT-based Osteoporosis Assessment with a Fully Automated Deep Learning Tool. Radiol. Artif. Intell. 2022, 4, e220042. [Google Scholar] [CrossRef] [PubMed]
  100. Zhao, Y.; Zhao, T.; Chen, S.; Zhang, X.; Sosa, M.S.; Liu, J.; Mo, X.; Chen, X.; Huang, M.; Li, S.; et al. Fully automated radiomic screening pipeline for osteoporosis and abnormal bone density with a deep learning-based segmentation using a short lumbar mDixon sequence. Quant. Imaging Med. Surg. 2022, 12, 1198–1213. [Google Scholar] [CrossRef] [PubMed]
  101. Hans, D.; Shevroja, E.; Leslie, W.D. Evolution in fracture risk assessment: Artificial versus augmented intelligence. Osteoporos. Int. 2021, 32, 209–212. [Google Scholar] [CrossRef] [PubMed]
  102. Yosibash, Z.; Trabelsi, N.; Buchnik, I.; Myers, K.W.; Salai, M.; Eshed, I.; Barash, Y.; Klang, E.; Tripto-Shkolnik, L. Hip Fracture Risk Assessment in Elderly and Diabetic Patients: Combining Autonomous Finite Element Analysis and Machine Learning. J. Bone Miner. Res. 2023, 38, 876–886. [Google Scholar] [CrossRef] [PubMed]
Figure 1. PRISMA flowchart for the literature search (adapted from the PRISMA group, 2020), which describes the selection of relevant articles.
Figure 1. PRISMA flowchart for the literature search (adapted from the PRISMA group, 2020), which describes the selection of relevant articles.
Bioengineering 11 00484 g001
Figure 3. Diagram showing the process of medical image data handling for machine learning.
Figure 3. Diagram showing the process of medical image data handling for machine learning.
Bioengineering 11 00484 g003
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Liu, R.W.; Ong, W.; Makmur, A.; Kumar, N.; Low, X.Z.; Shuliang, G.; Liang, T.Y.; Ting, D.F.K.; Tan, J.H.; Hallinan, J.T.P.D. Application of Artificial Intelligence Methods on Osteoporosis Classification with Radiographs—A Systematic Review. Bioengineering 2024, 11, 484. https://doi.org/10.3390/bioengineering11050484

AMA Style

Liu RW, Ong W, Makmur A, Kumar N, Low XZ, Shuliang G, Liang TY, Ting DFK, Tan JH, Hallinan JTPD. Application of Artificial Intelligence Methods on Osteoporosis Classification with Radiographs—A Systematic Review. Bioengineering. 2024; 11(5):484. https://doi.org/10.3390/bioengineering11050484

Chicago/Turabian Style

Liu, Ren Wei, Wilson Ong, Andrew Makmur, Naresh Kumar, Xi Zhen Low, Ge Shuliang, Tan Yi Liang, Dominic Fong Kuan Ting, Jiong Hao Tan, and James Thomas Patrick Decourcy Hallinan. 2024. "Application of Artificial Intelligence Methods on Osteoporosis Classification with Radiographs—A Systematic Review" Bioengineering 11, no. 5: 484. https://doi.org/10.3390/bioengineering11050484

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop