Next Article in Journal
Cascaded Spatial and Depth Attention UNet for Hippocampus Segmentation
Previous Article in Journal
A Hybrid Framework for Red Blood Cell Labeling Using Elliptical Fitting, Autoencoding, and Data Augmentation
Previous Article in Special Issue
The Contribution of AIDA (Artificial Intelligence Dystocia Algorithm) to Cesarean Section Within Robson Classification Group
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Evaluation of AI Performance in Spinal Radiographic Measurements Compared to Radiologists: A Study of Accuracy and Efficiency

1
Radiology Unit, Department of Surgical and Medical Sciences and Translational Medicine, Sapienza University of Rome, Sant’Andrea Hospital, 00189 Rome, Italy
2
Department of Biomedical Sciences, Humanitas University, 20100 Rozzano, Italy
3
Department of Diagnostic Imaging, IRCCS Humanitas, Research Hospital, 20100 Rozzano, Italy
*
Author to whom correspondence should be addressed.
J. Imaging 2025, 11(9), 310; https://doi.org/10.3390/jimaging11090310
Submission received: 19 June 2025 / Revised: 20 August 2025 / Accepted: 2 September 2025 / Published: 10 September 2025

Abstract

This study aimed to evaluate the reliability of an AI-based software tool in measuring spinal parameters—Cobb angle, thoracic kyphosis, lumbar lordosis, and pelvic obliquity—compared to manual measurements by radiologists and to assess potential time savings. In this retrospective monocentric study, 56 patients who underwent full-spine weight-bearing X-rays were analyzed. Measurements were independently performed by an experienced radiologist, a radiology resident, and the AI software. A consensus between two senior experts established the ground truth. Lin’s Concordance Correlation Coefficient (CCC), mean absolute error (MAE), ICC, and paired t-tests were used for statistical analysis. The AI software showed excellent agreement with human readers (CCC > 0.9) and demonstrated lower MAE than the resident in Cobb angle and lumbar lordosis measurements but slightly underperformed in thoracic kyphosis and pelvic obliquity. Importantly, the AI significantly reduced analysis time compared to both the experienced radiologist and the resident (p < 0.001). These findings suggest that the AI tool offers a reliable and time-efficient alternative to manual spinal measurements and may enhance accuracy for less experienced radiologists.

1. Introduction

Back pain is among the most frequent reasons for visiting a general practitioner or physiotherapist. Although magnetic resonance (MR) conducted in a supine position represents the gold standard for the detection of causes of back pain such as disk degeneration, Modic changes, and facet joint degeneration and offers many advantages in terms of soft tissue resolution and the absence of the use of ionizing radiation, it cannot accurately evaluate scoliosis, degree of thoracic kyphosis, lumbar lordosis, and pelvic obliquity. It is essential that spine X-rays are performed under weight-bearing conditions for the assessment of these parameters [1,2].
Manual measurements have several significant limitations. First, they are time-consuming, particularly when multiple measurements are required. Furthermore, their accuracy depends on the skills and experience of the operator, which can result in both inter-observer and intra-observer variability [3,4]. In this context, technological advancements have led to the increasing application of artificial intelligence (AI) in musculoskeletal (MSK) imaging [5,6,7,8]. AI is being increasingly adopted as a support tool to enhance the efficiency and accuracy of radiological assessments, providing fast, reliable, and reproducible measurements while minimizing errors caused by distraction and fatigue [9,10,11].
Recent advancements in artificial intelligence (AI), the capability of computational systems to perform tasks typically associated with human intelligence, such as learning, reasoning, perception, and decision-making, have opened up new frontiers in medical imaging, particularly in musculoskeletal radiology. AI-based tools promise to increase efficiency, reduce variability, and support less experienced radiologists in complex measurements. In this context, several algorithms have already been proposed for tasks such as automated Cobb angle quantification, hip morphology assessment, and spinopelvic parameter estimation. These models, often based on convolutional neural networks or U-Net–like segmentation pipelines, have shown promising accuracy and reproducibility across different clinical scenarios. However, growing evidence suggests that the integration of AI into clinical workflows does not always yield uniform benefits. A recent large-scale study by Harvard Medical School and collaborators [12] revealed that AI assistance may improve or impair diagnostic performance depending on the individual radiologist’s characteristics, highlighting the need for tailored AI implementation. In this context, evaluating the reliability, accuracy, and practical utility of AI tools in real-world radiological tasks is essential. The integration of AI into radiology education is gaining momentum, with growing interest among faculty, residents, and medical students. A recent survey by Hoyer et al. [13] highlighted the perceived value of AI-powered platforms for enhancing learning through interactive case-based training, automated feedback, and simulated reporting. While enthusiasm for AI in education is high, the study also revealed the need for user-centered design, transparency, and alignment with learners’ expectations. These findings underscore the importance of carefully structured AI integration—not only in clinical workflows but also in radiological training and assessment.
The primary aim of this study is to evaluate the reliability of an AI-based software tool, BoneMetrics (version 2.3.1, Gleamer, Paris, France), chosen because it is currently implemented in our institution and routinely used in clinical practice, in measurements of different spine parameters and to compare its accuracy to radiologists. Our secondary aim was to quantify time savings achieved using the automated software compared to traditional manual methods. The novelty of this is that, unlike prior works, we compared AI not only with an expert radiologist but also with a resident, thus exploring its role in supporting less experienced readers. Our main contributions are (i) validation of Bone Metrics in spine imaging, (ii) benchmarking against readers with different expertise, and (iii) demonstration of workflow efficiency gains. The study followed a simple retrospective framework with expert consensus as the reference standard and standard statistical analysis.

2. Materials and Methods

2.1. Patient Population

We conducted a retrospective monocentric study at Sant’Andrea Hospital, Sapienza University of Rome, and enrolled patients who underwent whole-spine weight-bearing X-rays from January 2023 to October 2024. Patients who refused to provide informed consent and images with severe artifacts were excluded from the study. Examinations were performed for different clinical indications: in pediatric patients, they were performed mainly for suspected scoliosis, while in adult patients, they were performed for degenerative spine diseases and suspected spinal instability. This study was approved by the IRB (approval code CE 6597/2021), and informed consent was obtained from all participants.

2.2. Image Acquisition

All images were acquired using the same digital radiography scanner (Bloomix 120 ED-R, Trade Art Manifacturing, Via della Pisana 1353, 00163 Rome, Italy) with patients in a standing weight-bearing position, in order to reproduce physiological load conditions on the spine. For each patient, both anteroposterior (AP) and lateral projections were obtained, covering the entire spine from the second cervical vertebra (C2) to the last sacral vertebral body (S5), including the pelvic bones and both femoral heads to allow evaluation of spinopelvic parameters. The dedicated scanner used in this study enables the acquisition of the entire spine in a single exposure, without the need for image stitching, thus reducing potential misalignment artifacts and ensuring greater measurement accuracy. Standardized positioning protocols were adopted to minimize patient rotation and to maintain consistent image quality across all examinations. All acquisitions were performed by experienced radiographers according to the department’s routine protocol for whole-spine evaluation. Exposure parameters, such as tube voltage and current, were adjusted based on patient size and age to optimize image quality while maintaining the radiation dose as low as reasonably achievable (ALARA principle) (Figure 1).

2.3. Image Analysis

Image analysis was performed independently by a radiologist with 25 years of experience in musculoskeletal (MSK) radiology (GA) and by a radiology resident (GG) with 3 years of experience. Subsequently, the images were also analyzed by the AI software. The AI-based software, BoneMetrics (version 2.3.1, Gleamer, Paris, France), is a CE-certified AI-based software tool designed to automatically perform musculoskeletal measurements on conventional radiographs and EOS acquisitions. The system integrates several convolutional neural network models for the detection and localization of anatomical landmarks, from which quantitative parameters are subsequently derived. For each landmark, the algorithm provides a confidence score ranging from 0 to 100; only points exceeding a threshold of 50 are included in the final measurement process. The software architecture combines different approaches, including a top–down framework based on Detectron2, a lightweight high-resolution network (LiteHRNet), and a bottom–up strategy. This multi-architecture design enhances the adaptability and robustness of the model across diverse radiographic datasets. The training dataset comprised more than 5000 radiographs collected from over 20 European centers, including cases both with and without orthopedic implants. Annotation was performed by radiographers and radiologists with specific training and subsequently reviewed by an experienced musculoskeletal radiologist with over 14 years of expertise to ensure data quality. Importantly, none of the radiographs or patients included in the current study were used in the training phase of the software [14].
The radiologists were blinded to the measurements obtained by the AI system. The following measurements were performed: Cobb angle (defined as the angle formed by lines drawn parallel to the endplates of the two vertebrae that exhibit the greatest inclination relative to the horizontal axis) and pelvic obliquity (determined as the distance between the pelvic coronal reference line and a horizontal line aligned parallel to the floor, measured on the AP view); thoracic kyphosis (measured as the angle formed by two lines passing through the superior endplate of T1 and the inferior endplate of T12); and lumbar lordosis (measured as the angle formed by two lines passing through the superior endplate of L1 and the inferior endplate of L5), calculated on the lateral view.
All measurements were performed directly with the PACS system (GE Healthcare, Waukesha, WI, USA) using the standard built-in tools for measuring angles and distances, without the use of dedicated post-processing workstations or advanced image analysis software [12,13].
For validation of the measurements obtained by both radiologists and the AI tool, an orthopedic surgeon with 35 years of experience in spine surgery (NM) and another radiologist with 15 years of experience in musculoskeletal (MSK) radiology (RP) were considered as the reference standard (ground truth).
In addition, the time required to perform the measurements was recorded for both human readers and the AI software, in order to assess the potential impact on workflow efficiency.

2.4. Statistical Analysis

Statistical analyses were performed using MedCalc (MedCalc Software, version15, Ostend, Belgium). Data were reported as mean ± standard deviation (SD) or as absolute differences, depending on the analysis type. Paired t-tests were used to compare measurements between the radiologists and AI. Separate analyses were conducted for each parameter, including Cobb angle, thoracic kyphosis, lumbar lordosis, and pelvic obliquity.
The agreement between the experienced radiologist, radiology resident, and AI tool was assessed using Lin’s Concordance Correlation Coefficient (CCC). CCC values greater than 0.90 were interpreted as showing excellent agreement. In addition, inter- and intra-observer reliability were assessed by calculating the Intraclass Correlation Coefficient (ICC, two-way random effects, absolute agreement). ICC values above 0.90 were considered indicative of excellent reliability.
The accuracy of the resident and AI was evaluated by calculating the Mean Absolute Error (MAE) relative to the expert’s measurements, which were used as the reference standard.
The statistical significance threshold was set at p < 0.05.

3. Results

3.1. Patient Population

From an initial population of 67 patients, a total of 56 were included in the study (21 males, 35 females). The participants had a mean age of 33.65 ± 24.51 years (range 7–74 years). Among the included subjects, 23 were underage patients. Images with severe artifacts (n = 9) and patients who refused to provide informed consent (n = 2) were excluded from the analysis. Pathological findings were detected in 49/56 patients (87.5%). Specifically, scoliosis was found in 16/56 patients (28.6%), spondylolisthesis was found in 4/56 patients (7.1%), and a vertebral collapse was identified in 1/56 patient (1.8%) (Figure 2).

3.2. Image Acquisition

A total of 122 projections were acquired. In six patients, additional projections were required due to incorrect positioning, which initially prevented accurate measurements. For full-spine acquisitions, correct shoulder positioning in the lateral (LL) view was essential to avoid superimposition of the cervico-dorsal junction, a key reference for evaluating vertical axis alignment. In the anteroposterior (AP) view, the mean tube voltage was 82 kV (range 80–90 kV), with a tube current of 200 mA, an exposure of 64 mAs, and a time of 320 ms. In the lateral view, the mean tube voltage was 84 kV (range 80–90 kV), with a tube current of 200 mA, an exposure of 80 mAs, and the same acquisition time of 320 ms.

3.3. Image Analysis

No significant differences were found in any of the measurements between AI and the experienced radiologist or the radiology resident (all p > 0.05). Likewise, no differences emerged between the two human readers (all p > 0.05). Finally, no statistically significant differences were observed between pediatric and adult patient groups (p > 0.05) (Table 1).
The concordance between measurements by the experienced radiologist, radiology resident, and AI demonstrated a high level of agreement across all parameters, with CCC values exceeding 0.9 in most cases (Figure 3 and Figure 4).
For the Cobb angle, the AI achieved a lower MAE (0.577) compared to the resident (0.632), indicating slightly greater accuracy. For thoracic kyphosis, the resident exhibited a lower MAE (2.375) compared to the AI (2.464), suggesting better agreement with the expert in this measurement. In the case of lumbar lordosis, the AI demonstrated superior accuracy with a lower MAE (1.395) compared to the resident (1.902). Conversely, for pelvic obliquity, the resident achieved a lower MAE (0.296) compared to the AI (0.475) (Table 2). The inter-observer agreement between the experienced radiologist and the resident was excellent for all parameters (ICC range 0.905–0.996). When including the AI software as a third reader, the overall concordance remained excellent (ICC range 0.931–0.996). In addition, intra-observer reproducibility was also excellent across all parameters (ICC > 0.90).
Analysis time was significantly shorter for AI vs. the experienced radiologist (50.00 ± 0.00 s vs. 147.34 ± 18.78 s, p < 0.001) and for AI vs. the radiology resident (50.00 ± 0.00 s vs. 231.78 ± 20.15, p < 0.001). Analysis was significantly faster for the experienced radiologist vs. the radiology resident (147.34 ± 18.78 s vs. 231.78 ± 20.15, p = 0.027) (Table 3) (Figure 5).

4. Discussion

Our study demonstrated that the AI tool BoneMetrics (version2.3.1, Gleamer, Paris, France) has the same accuracy as an experienced radiologist and radiology resident in measuring Cobb angle, pelvic obliquity, thoracic kyphosis, and lumbar lordosis on a weight bearing spine X-ray with a significant analysis time reduction. Furthermore, if not significant, a p-value (0.062) for Cobb angle measurements between the AI tool and radiology resident suggest that this tool could help non-experienced radiologists with this measurement. The AI system performed better than the resident in some parameters, such as the Cobb angle and Lumbar Lordosis, while the resident showed better performance in others, such as Thoracic Kyphosis and Pelvic Obliquity. These results suggest that leveraging both AI and human expertise could optimize accuracy and reliability in radiological assessments.
Up to now, this study is the first to test this specific AI tool on the spine. Lassalle et al. [14,15] in their multicentric study demonstrated a comparable performance of this AI tool on weight-bearing forefoot and lateral foot radiographs, with relevant time savings. Similarly to our work, their study was retrospective and based on senior radiologists’ annotations as the ground truth, with only slightly more patients included. Unlike their study, however, our results highlight the additional value of this tool in supporting less experienced radiologists. Suri A. et colleagues [16], in their retrospective study, with experienced radiologists as the ground truth, presented a deep learning algorithm (SpineTK) trained with a fivefold cross-validation scheme and hyperparameter tuning to avoid overfitting, which is also available as open-source code. Their model demonstrated a strong agreement with manually annotated Cobb angle measurements performed by experienced radiologists and was statistically invariant across several clinical characteristics (hardware presence, sex, BMI, scoliosis severity, age, and image type). Moreover, they reported >90% accuracy in detecting scoliosis across all severity grades, and the algorithm proved robust in challenging clinical conditions such as severe scoliosis, different imaging positions, and the presence of extensive spinal hardware. Unlike previous studies such as Li et al. [17], where an AI algorithm (a multi-stage CNN-based pipeline consisting of spine region detection, vertebral segmentation with a U-Net-like architecture, centerline extraction, and automatic identification of the most tilted vertebrae) was used to automate Cobb angle measurements exclusively on the coronal plane, our study evaluated spinal alignment on both the sagittal and coronal planes. In the work by Li et al., the automated method showed a mean absolute deviation of 4.17° from expert readers, which was even lower than the average intra-reader variability (5.16°). However, their analysis was limited to scoliosis assessment alone. In contrast, our approach incorporates both coronal Cobb angle (scoliosis) and sagittal curvature metrics (e.g., thoracic kyphosis and lumbar lordosis), offering a more comprehensive characterization of spinal deformities [17].
Boel F. and colleagues [18], in their single-center study, investigated the agreement and reliability of manual and automated morphological measurements on AP pelvic radiographs for different measurements prospectively conducted in a population of 30 patients, and the mean of measurements performed by experienced orthopedic surgeons was used as the reference standard. Their presented algorithm (an automated pipeline based on BoneFinder® software, v.1.3.4 for landmark detection and a Python-based framework for radiographic measurement derivation) performed equally well compared to the current best practice of manual measurement by trained readers. By automatically annotating 80 anatomical landmarks on AP pelvic radiographs and subsequently deriving hip morphology indices, the system demonstrated both reliability and efficiency, while significantly reducing the time required for radiological measurements.
The performance of an algorithm developed with deep-learning and computer vision techniques, using a modified U-Net architecture with a ResNet34 backbone trained on more than 2900 pelvic radiographs for measurement of the lateral center edge angle of Wiberg (LCEA) and Acetabular index angle (AIA) when reading pelvic radiographs, was investigated by Jensen J. et al. [19]. Like our work, their study was retrospective, with only slightly more patients included (71); the lack of a ground truth against which the algorithm and human measurements could be compared was also a limitation for them. The manual identification of the same landmarks performed by five human readers was subject to variance, and the level of agreement between the algorithm and human readers was consequently poor, although a tendency was revealed for the senior orthopedic surgeon to agree the most with the algorithm, particularly for LCEA right measurements. Our results, compared to those obtained in these other studies, have shown the same performance for AI tools and experienced radiologists. The strength of our study is in the potential role of this AI tool to help non-expert MSK radiologists.
Our results are consistent with recent evidence by Harake et al. [20], who developed SpinePose, which employs three parallel CNNs: two top–down region-based models for L1 and S1 and a bottom–up encoder–decoder model for the remaining key points. This architecture enables precise localization of anatomical landmarks by combining region-specific detection with direct key point prediction on the native image. In their study, the model was trained on 761 lateral radiographs and demonstrated mean absolute errors (MAEs) ranging from 3° to 5° across parameters such as pelvic incidence, pelvic tilt, sacral slope, and lumbar lordosis. Furthermore, coefficient of determination (R2) values ranged between 0.85 and 0.92, confirming high consistency with manual annotations performed by experienced spine surgeons and neuroradiologists. These findings reinforce the idea that AI is not only a valuable tool for radiologists but can also support a broader spectrum of specialists, including orthopedic surgeons and physiatrists, particularly in the context of preoperative planning and spinal alignment evaluation.
Beyond quantitative performance metrics, recent evidence also highlights a growing positive perception of AI among radiologists. For example, in the prospective study by Hoppe et al. [21], the integration of AI tools not only improved diagnostic workflow but also enhanced radiologists’ confidence and acceptance of AI in clinical practice. Participants reported increased trust in the technology and acknowledged its utility as a support tool, especially in high-pressure settings. Similarly, in our study, the AI tool was not only accurate and time-efficient but also well-received by radiologists during implementation, suggesting that the integration of AI into spinal imaging can be perceived as a helpful and reliable adjunct rather than a threat to professional expertise. These findings support a broader acceptance of AI in radiology, where clinical utility, efficiency, and user perception are all essential in real-world adoption.
While our study, along with others, demonstrates the quantitative benefits of AI tools in terms of accuracy and efficiency, the integration of artificial intelligence into clinical radiology workflows also requires careful consideration of practical challenges, as highlighted by the recent literature. For instance, in their study, Korfiatis et al. [22] emphasize issues such as data standardization, integration into existing PACS/RIS systems, regulatory compliance, and the importance of clinician trust and training. These considerations are crucial in ensuring the real-world adoption and scalability of AI technologies.
The retrospective nature of the study and the small number of participants could be considered to represent the first limitation of the study. Furthermore, the monocentric nature could be considered a second potential limitation. A third limitation to the current study is the lack of an objective ground truth against which the algorithm and human measurements can be compared. Establishing a ground truth is, however, a common challenge in radiographic measurements. A fourth limitation is that the study was restricted to BoneMetrics, as it is the only AI tool currently implemented in our institution.

5. Conclusions

In conclusion, BoneMetrics demonstrated comparable accuracy to both an experienced radiologist and a radiology resident in measuring Cobb angle, pelvic obliquity, thoracic kyphosis, and lumbar lordosis on weight-bearing spine radiographs, while significantly reducing the analysis time. These results suggest that AI-based tools can provide a reliable and time-efficient alternative for routine clinical practice. Importantly, the use of BoneMetrics may be particularly valuable in supporting less experienced radiologists, ensuring consistency and reducing inter-observer variability. Future studies with larger and more diverse patient populations will be essential to further validate these findings and to explore the broader integration of AI into musculoskeletal imaging workflows.

Author Contributions

Conceptualization, F.P.; methodology, F.P. and M.Z.; software, D.D.S. and G.G.S.; validation, D.C. and A.L.; formal analysis, A.D.G., B.M., and G.T.; investigation, G.A. and R.P.; resources, T.P. and M.F.; data curation, M.P.; writing—original draft preparation, F.P. and G.G.S.; writing—review and editing, D.D.S., M.Z., and G.T.; visualization, R.P.; supervision, D.C. and A.L.; project administration, F.P., D.C., and A.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki and approved by the Institutional Review Board of Comitato Etico dell’Ospedale Sant’Andrea, Sapienza Università di Roma (protocol code CE 6597/2021, 12 January 2022).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AIArtificial Intelligence
MSKMusculoskeletal
MRMagnetic Resonance
APAnteroposterior
LLLatero-lateral (Lateral View)
CCCConcordance Correlation Coefficient
MAEMean Absolute Error
IRBInstitutional Review Board

References

  1. Zerunian, M.; Pucciarelli, F.; Caruso, D.; De Santis, D.; Polici, M.; Masci, B.; Nacci, I.; Del Gaudio, A.; Argento, G.; Redler, A.; et al. Fast high-quality MRI protocol of the lumbar spine with deep learning-based algorithm: An image quality and scanning time comparison with standard protocol. Skelet. Radiol. 2024, 53, 151–159. [Google Scholar] [CrossRef]
  2. Almahmoud, O.H.; Baniodeh, B.; Musleh, R.; Asmar, S.; Zyada, M.; Qattousah, H. Overview of adolescent idiopathic scoliosis and associated factors: A scoping review. Int. J. Adolesc. Med. Health 2023, 35, 437–441. [Google Scholar] [CrossRef]
  3. Kim, Y.-T.; Jeong, T.S.; Kim, Y.J.; Kim, W.S.; Kim, K.G.; Yee, G.T. Automatic Spine Segmentation and Parameter Measurement for Radiological Analysis of Whole-Spine Lateral Radiographs Using Deep Learning and Computer Vision. J. Digit. Imaging 2023, 36, 1447–1459. [Google Scholar] [CrossRef]
  4. Hussain, D.; Han, S.-M. Computer-aided osteoporosis detection from DXA imaging. Comput. Methods Programs Biomed. 2019, 173, 87–107. [Google Scholar] [CrossRef] [PubMed]
  5. Harvey, H.B.; Gowda, V. Clinical applications of AI in MSK imaging: A liability perspective. Skelet. Radiol. 2022, 51, 235–238. [Google Scholar] [CrossRef]
  6. Huber, F.A.; Guggenberger, R. AI MSK clinical applications: Spine imaging. Skelet. Radiol. 2022, 51, 279–291. [Google Scholar] [CrossRef] [PubMed]
  7. Joseph, G.B.; McCulloch, C.E.; Sohn, J.H.; Pedoia, V.; Majumdar, S.; Link, T.M. AI MSK clinical applications: Cartilage and osteoarthritis. Skelet. Radiol. 2022, 51, 331–343. [Google Scholar] [CrossRef] [PubMed]
  8. Yi, P.H.; Mutasa, S.; Fritz, J. AI MSK clinical applications: Orthopedic implants. Skelet. Radiol. 2022, 51, 305–313. [Google Scholar] [CrossRef] [PubMed]
  9. Ruitenbeek, H.C.; Oei, E.H.G.; Visser, J.J.; Kijowski, R. Artificial intelligence in musculoskeletal imaging: Realistic clinical applications in the next decade. Skelet. Radiol. 2024, 53, 1849–1868. [Google Scholar] [CrossRef]
  10. Alowais, S.A.; Alghamdi, S.S.; Alsuhebany, N.; Alqahtani, T.; Alshaya, A.I.; Almohareb, S.N.; Aldairem, A.; Alrashed, M.; Bin Saleh, K.; Badreldin, H.A.; et al. Revolutionizing healthcare: The role of artificial intelligence in clinical practice. BMC Med. Educ. 2023, 23, 689. [Google Scholar] [CrossRef]
  11. Novak, A.; Hollowday, M.; Morgado, A.T.E.; Oke, J.; Shelmerdine, S.; Woznitza, N.; Metcalfe, D.; Costa, M.L.; Wilson, S.; Kiam, J.S.; et al. Evaluating the impact of artificial intelligence-assisted image analysis on the diagnostic accuracy of front-line clinicians in detecting fractures on plain X-rays (FRACT-AI): Protocol for a prospective observational study. BMJ Open 2024, 14, e086061. [Google Scholar] [CrossRef]
  12. Does AI Help or Hurt Human Radiologists’ Performance? It Depends on the Doctor|Harvard Medical School. 2024. Available online: https://hms.harvard.edu/news/does-ai-help-or-hurt-human-radiologists-performance-depends-doctor (accessed on 26 June 2025).
  13. Li, R.; Liu, G.; Zhang, M.; Rong, D.; Su, Z.; Shan, Y.; Lu, J. Integration of artificial intelligence in radiology education: A requirements survey and recommendations from faculty radiologists, residents, and medical students. BMC Med. Educ. 2025, 25, 380. [Google Scholar] [CrossRef]
  14. Lassalle, L.; Regnard, N.E.; Durteste, M.; Ventre, J.; Marty, V.; Clovis, L.; Zhang, Z.; Nitche, N.; Ducarouge, A.; Laredo, J.-D.; et al. Evaluation of a deep learning software for automated measurements on full-leg standing radiographs. Knee Surg. Relat. Res. 2024, 36, 40. [Google Scholar] [CrossRef]
  15. Lassalle, L.; Regnard, N.-E.; Ventre, J.; Marty, V.; Clovis, L.; Zhang, Z.; Nitche, N.; Guermazi, A.; Laredo, J.-D. Automated weight-bearing foot measurements using an artificial intelligence–based software. Skelet. Radiol. 2025, 54, 229–241. [Google Scholar] [CrossRef]
  16. Suri, A.; Tang, S.; Kargilis, D.; Taratuta, E.; Kneeland, B.J.; Choi, G.; Agarwal, A.; Anabaraonye, N.; Xu, W.; Parente, J.B.; et al. Conquering the Cobb Angle: A Deep Learning Algorithm for Automated, Hardware-Invariant Measurement of Cobb Angle on Radiographs in Patients with Scoliosis. Radiol. Artif. Intell. 2023, 5, e220158. [Google Scholar] [CrossRef]
  17. Li, K.; Gu, H.; Colglazier, R.; Lark, R.; Hubbard, E.; French, R.; Smith, D.; Zhang, J.; McCrum, E.; Catanzano, A.; et al. Deep learning automates Cobb angle measurement compared with multi-expert observers. arXiv 2024, arXiv:2403.12115. [Google Scholar] [CrossRef]
  18. Boel, F.; de Vos-Jakobs, S.; Riedstra, N.; Lindner, C.; Runhaar, J.; Bierma-Zeinstra, S.; Agricola, R. Automated radiographic hip morphology measurements: An open-access method. Osteoarthr. Imaging 2024, 4, 100181. [Google Scholar] [CrossRef]
  19. Jensen, J.; Graumann, O.; Overgaard, S.; Gerke, O.; Lundemann, M.; Haubro, M.H.; Varnum, C.; Bak, L.; Rasmussen, J.; Olsen, L.B.; et al. A Deep Learning Algorithm for Radiographic Measurements of the Hip in Adults—A Reliability and Agreement Study. Diagnostics 2022, 12, 2597. [Google Scholar] [CrossRef] [PubMed]
  20. Harake, E.S.; Linzey, J.R.; Jiang, C.; Joshi, R.S.; Zaki, M.M.; Jones, J.C.; Khalsa, S.S.S.; Lee, J.H.; Wilseck, Z.; Joseph, J.R.; et al. Development and validation of an artificial intelligence model to accurately predict spinopelvic parameters. J. Neurosurg. Spine 2024, 41, 88–96. [Google Scholar] [CrossRef]
  21. Hoppe, B.F.; Rueckel, J.; Dikhtyar, Y.; Heimer, M.; Fink, N.; Sabel, B.O.; Ricke, J.; Rudolph, J.; Cyran, C.C. Implementing Artificial Intelligence for Emergency Radiology Impacts Physicians’ Knowledge and Perception: A Prospective Pre- and Post-Analysis. Investig. Radiol. 2024, 59, 404–412. [Google Scholar] [CrossRef] [PubMed]
  22. Korfiatis, P.; Kline, T.L.; Meyer, H.M.; Khalid, S.; Leiner, T.; Loufek, B.T.; Blezek, D.; Vidal, D.E.; Hartman, R.P.; Joppa, L.J.; et al. Implementing Artificial Intelligence Algorithms in the Radiology Workflow: Challenges and Considerations. Mayo Clin. Proc. Digit. Health 2025, 3, 100188. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Full-spine radiographs were acquired in standing position using the same dedicated scanner (Bloomix 120 ED-R). (A) Anteroposterior (AP) view; (B) lateral view. For each patient, AP and lateral projections were obtained from the base of the skull to the last sacral segment, including the hip bones and femoral heads. The scanner enabled the acquisition of the entire spine in a single shot, without the need for image stitching.
Figure 1. Full-spine radiographs were acquired in standing position using the same dedicated scanner (Bloomix 120 ED-R). (A) Anteroposterior (AP) view; (B) lateral view. For each patient, AP and lateral projections were obtained from the base of the skull to the last sacral segment, including the hip bones and femoral heads. The scanner enabled the acquisition of the entire spine in a single shot, without the need for image stitching.
Jimaging 11 00310 g001
Figure 2. Patient enrolment flow-chart.
Figure 2. Patient enrolment flow-chart.
Jimaging 11 00310 g002
Figure 3. Image analysis of Cobb angle and pelvic obliquity in the same patients. (A) Automated measurements; (B) radiologist measurements.
Figure 3. Image analysis of Cobb angle and pelvic obliquity in the same patients. (A) Automated measurements; (B) radiologist measurements.
Jimaging 11 00310 g003
Figure 4. Sagittal spine image analysis in the same patient. (A) Automated measurements of thoracic kyphosis (T4–T12) and lumbar lordosis (L1–S1); (B) manual measurements performed by a radiologist.
Figure 4. Sagittal spine image analysis in the same patient. (A) Automated measurements of thoracic kyphosis (T4–T12) and lumbar lordosis (L1–S1); (B) manual measurements performed by a radiologist.
Jimaging 11 00310 g004
Figure 5. Comparison of mean analysis time (seconds) among artificial intelligence (AI) software, the experienced radiologist, and the radiology resident. AI software required significantly shorter analysis times compared to both human readers, while the experienced radiologist outperformed the resident. Error bars represent standard deviations.
Figure 5. Comparison of mean analysis time (seconds) among artificial intelligence (AI) software, the experienced radiologist, and the radiology resident. AI software required significantly shorter analysis times compared to both human readers, while the experienced radiologist outperformed the resident. Error bars represent standard deviations.
Jimaging 11 00310 g005
Table 1. Mean values (±standard deviation) for Cobb angle, pelvic obliquity, lumbar lordosis, and thoracic kyphosis as evaluated by the radiologist, the resident, and artificial intelligence software.
Table 1. Mean values (±standard deviation) for Cobb angle, pelvic obliquity, lumbar lordosis, and thoracic kyphosis as evaluated by the radiologist, the resident, and artificial intelligence software.
Cobb Angle (°)p ValuePelvic Obliquity (mm)p ValueThoracic Kyphosis (°)p ValueLumbar Lordosis (°)p Value
AI vs. Experienced Radiologist12.75 ± 11.460.1102.49 ± 1.080.57238.42 ± 11.470.33658.25 ± 11.740.282
AI vs. Radiology Resident12.75 ± 11.460.0622.49 ± 1.080.33638.42 ± 11.470.06258.25 ± 11.740.135
Experienced Radiologist vs. Radiology Resident12.57 ± 11.270.3562.53 ± 1.060.56238.04 ± 11.810.12757.83 ± 11.400.265
Table 2. Comparison of mean absolute error (MAE) and Concordance Correlation Coefficient (CCC) between the expert, the resident, and artificial intelligence (AI) software for the evaluation of Cobb angle, thoracic kyphosis, lumbar lordosis, and pelvic obliquity.
Table 2. Comparison of mean absolute error (MAE) and Concordance Correlation Coefficient (CCC) between the expert, the resident, and artificial intelligence (AI) software for the evaluation of Cobb angle, thoracic kyphosis, lumbar lordosis, and pelvic obliquity.
MAE (Resident)MAE (AI)CCC
(Expert–Trainee)
CCC (Expert–AI)CCC (Trainee–AI)
COBB ANGLE0.6320.5770.9960.9970.994
THORACIC KYPHOSIS 2.3752.4640.9040.9130.989
LUMBAR LORDOSIS1.9021.3950.9090.9840.898
PELVIC OBLIQUITY 0.2960.4750.9940.9880.988
Table 3. Comparison of analysis time (seconds) between the artificial intelligence (AI) software, experienced radiologist, and radiology resident.
Table 3. Comparison of analysis time (seconds) between the artificial intelligence (AI) software, experienced radiologist, and radiology resident.
Analysis Time (Seconds)p-Value
AI vs. Experienced Radiologist50.00 ± 0.00147.34 ± 18.78p < 0.001
AI vs. Radiology Resident50.00 ± 0.00231.78 ± 20.15p < 0.001
Experienced Radiologist vs. Radiology Resident147.34 ± 18.78231.78 ± 20.15p = 0.027
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Pucciarelli, F.; Gentiloni Silveri, G.; Zerunian, M.; De Santis, D.; Polici, M.; Del Gaudio, A.; Masci, B.; Polidori, T.; Tremamunno, G.; Persechino, R.; et al. Evaluation of AI Performance in Spinal Radiographic Measurements Compared to Radiologists: A Study of Accuracy and Efficiency. J. Imaging 2025, 11, 310. https://doi.org/10.3390/jimaging11090310

AMA Style

Pucciarelli F, Gentiloni Silveri G, Zerunian M, De Santis D, Polici M, Del Gaudio A, Masci B, Polidori T, Tremamunno G, Persechino R, et al. Evaluation of AI Performance in Spinal Radiographic Measurements Compared to Radiologists: A Study of Accuracy and Efficiency. Journal of Imaging. 2025; 11(9):310. https://doi.org/10.3390/jimaging11090310

Chicago/Turabian Style

Pucciarelli, Francesco, Guido Gentiloni Silveri, Marta Zerunian, Domenico De Santis, Michela Polici, Antonella Del Gaudio, Benedetta Masci, Tiziano Polidori, Giuseppe Tremamunno, Raffaello Persechino, and et al. 2025. "Evaluation of AI Performance in Spinal Radiographic Measurements Compared to Radiologists: A Study of Accuracy and Efficiency" Journal of Imaging 11, no. 9: 310. https://doi.org/10.3390/jimaging11090310

APA Style

Pucciarelli, F., Gentiloni Silveri, G., Zerunian, M., De Santis, D., Polici, M., Del Gaudio, A., Masci, B., Polidori, T., Tremamunno, G., Persechino, R., Argento, G., Francone, M., Laghi, A., & Caruso, D. (2025). Evaluation of AI Performance in Spinal Radiographic Measurements Compared to Radiologists: A Study of Accuracy and Efficiency. Journal of Imaging, 11(9), 310. https://doi.org/10.3390/jimaging11090310

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop