Next Article in Journal
Echo Intensity Correction Method for Ultrasound Computed Tomography in Musculoskeletal Imaging
Next Article in Special Issue
Depth-Dependent Performance of Residual Networks for Low-Count PET Image Restoration Using a Dedicated 3D-Printed Striatum Phantom
Previous Article in Journal
Surface EMG-Validated Multi-DoF Wheelchair-Based Rehabilitation Device
Previous Article in Special Issue
Simulation-Driven Annotation-Free Deep Learning for Automated Detection and Segmentation of Airway Mucus Plugs on Non-Contrast CT Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Improved AI-Assisted Image Recognition of Cervical Spine Vertebrae Enables Motion Pattern Analysis in Dynamic X-Ray Recordings

by
Esther van Santbrink
1,2,3,*,
Tijmen H. W. Hijzelaar
4,
Valérie N. E. Schuermans
1,2,3,
Anouk Y. J. M. Smeets
1,2,
Henk van Santbrink
1,2,3,
Rob de Bie
3,5,
Mitko Veta
4 and
Toon F. M. Boselie
1,2,3
1
Department of Neurosurgery, Maastricht University Medical Centre, 6229 HX Maastricht, The Netherlands
2
Department of Neurosurgery, Zuyderland Medical Centre, 6419 PC Heerlen, The Netherlands
3
CAPHRI Institute for Public Health and Primary Care, Maastricht University, 6211 LK Maastricht, The Netherlands
4
Department of Biomedical Engineering, Eindhoven University of Technology, 5612 AZ Eindhoven, The Netherlands
5
Department of Epidemiology, Maastricht University, 6211 LK Maastricht, The Netherlands
*
Author to whom correspondence should be addressed.
Bioengineering 2026, 13(3), 351; https://doi.org/10.3390/bioengineering13030351
Submission received: 1 February 2026 / Revised: 2 March 2026 / Accepted: 11 March 2026 / Published: 18 March 2026
(This article belongs to the Special Issue Artificial Intelligence-Based Medical Imaging Processing)

Abstract

Background: Qualitative motion analysis revealed that the cervical spine moves according to a consistent pattern. Current data analysis methods are limited by the extensive time required to process the retrieved data. A previous study demonstrated the feasibility of using a deep-learning model to automate analysis methods. However, segmentation accuracy needed to be improved. This study aims to improve segmentation model performance to enable reliable motion analysis. Methods: Four nnU-Net configurations were tested: baseline (A), pre-trained (B), with histogram equalization (C), and pre-trained with histogram equalization (D). Segmentation performance was evaluated using Dice Similarity Coefficient (DSC), Intersection over Union (IoU) and 95th percentile Hausdorff Distance (HD95). Vertebral rotation was estimated using mean shapes. Reliability was assessed using the Intraclass Correlation Coefficient (ICC). Sensitivity analyses were conducted. Results: Across all models, mean DSC ranged from 0.67 to 0.92, mean IoU from 0.55 to 0.85, and mean HD95 from 2.35 to 19.67 mm. After sensitivity analysis for low segmental range of motion (sROM) and low-quality recordings, the mean ICC ranged from 0.617 to 0.837 for model A, from 0.609 to 0.780 for model B, from 0.409 to 0.689 for model C, and from 0.480 to 0.835 for model D. Conclusions: This study shows that Models A and B can accurately analyze cervical motion patterns. High image contrast and an adequate sROM are essential for robust model performance. It also marks an important step toward automated qualitative motion analysis, increasing the accessibility of motion pattern evaluation.

1. Introduction

Qualitative motion analysis of the cervical spine revealed that a consistent motion pattern is present in 80% to 90% of healthy young individuals [1]. This pattern is defined as the order of maximum contribution of each motion segment to the extension movement. The consistent motion pattern is uncovered by the sequence of segmental contribution (SSC), which is determined by tracking the relative rotation of cervical vertebrae throughout dynamic X-ray recordings [2]. The SSC is a reliable metric, with a mean sensitivity of 90% and specificity of 85% for differentiating motion patterns between asymptomatic individuals and patients with radicular syndrome [1].
An RCT comparing anterior cervical discectomy with arthroplasty (ACDA) to anterior cervical discectomy (ACD) in patients with a radicular syndrome demonstrated that a motion pattern consistent with that of asymptomatic controls was present in 16.6% of ACDA patients before surgery. One year postoperatively, this pattern was preserved or restored in 66.7% of the ACDA group [3,4]. After a mean follow-up of 11 years, the predefined motion pattern was absent in these patients, although segmental range of motion (sROM) of the index level and upper adjacent level were maintained. The timing and underlying mechanisms responsible for this absence are unknown. In elderly individuals, quality of motion decreases over time due to the natural aging process. Qualitative analysis of cervical motion in healthy elderly individuals showed that the previously defined consistent motion pattern was absent [5]. A more comprehensive understanding how motion patterns evolve is required to define normal motion. In clinical contexts, identifying a motion pattern could influence decision-making processes, such as the selection of a surgical approach.
Qualitative motion analysis has not been widely adopted in research or clinical settings because it is labor intensive and requires trained experts. Artificial intelligence (AI) models can provide automatic and quantitative analysis of the motion pattern, thus eliminating both workload and subjectivity constraints. Several architectures for AI models have been explored to analyze medical imaging data, including 2D, 2D with temporal information (2D+t), and 3D approaches. Their segmentation performance is typically evaluated using overlap-based metrics such as the Dice Similarity Coefficient (DSC), which measures the overlap between the predicted segmentation and the ground truth. Manually annotated images are often considered the ground truth.
Previous work demonstrated the feasibility of using deep learning to analyze cervical spine motion in dynamic X-ray recordings [6]. Segmentation performance showed low to high accuracy, with Intersection over Union (IoU) ranging from 0.37 to 0.74 and DSC from 0.53 to 0.80. Reliability of motion estimation, assessed by Intraclass Correlation Coefficient (ICC), ranged from 0.62 to 0.96 for individual vertebrae and from 0.28 to 0.72 for vertebral segments, with higher scores observed in 2D+t models. These findings highlight the potential of motion pattern analysis, though segmentation performance needs to be improved to support routine application in clinical research and decision-making.
To improve segmentation accuracy, we trained a model based on nnU-Net, which is a state-of-the-art medical image segmentation model [7]. This model is trained automatically, with model training hyperparameters optimized for best performance, ensuring great efficiency. The model was originally developed using 23 biomedical segmentation datasets.
Among the different methods available to improve the performance of nnU-net, this study focuses on pre-training and histogram equalization due to their specific suitability for our dataset. Transfer learning via pre-training uses pre-trained models to mitigate data scarcity [8]. Histogram equalization augments the contrast of input images to further refine model performance [9]. This could potentially remedy the poor segmentation performance of the vertebrae that lack contrast due to overprojection of other structures, such as vertebrae C6 and C7.
This study aimed to test and improve nnU-net model performances in cervical dynamic X-ray recordings to enable reliable motion analysis in research and clinical settings.

2. Materials and Methods

2.1. Population and Manual Annotation

Previous studies by our group investigated motion patterns in healthy, asymptomatic individuals in different age groups, and in patients with a cervical radicular syndrome before and after surgery [1,3,5,10,11]. Manually annotated data from these studies were combined to train the models. Dynamic X-ray recordings were all made following the same protocol. Participants sat on a height-adjustable crutch with neck, shoulders, and head free. They moved their head fluently from full extension to full flexion and back in about 10 s, guided by a metronome. The upper body remained still throughout the movement. Shoulders and arms were kept low to ensure all cervical vertebrae remained visible during recording.
Template areas containing each vertebral body and the skull base were manually drawn by two trained individuals (TB and VS) in the median frame of the recording, labeled C0–C7. The bottom part of the sella turcica and clivus are projected in the X-ray image as a stable and hyperintense structure of the skull and serve as an ideal template to track the movement of C0 (the skull base). Image recognition software, based on matching normalized gradient field images on a best-fit principle, was used to locate the vertebrae in the other frames of the recordings [12]. The contours were manually checked and adjusted in every frame when necessary. These annotations were considered the ground truth, and previous studies have reported excellent intra- and interobserver reliability [1,5].

2.2. Dataset

The dataset included a total of 112 recordings, of which 89 recordings were from 39 healthy individuals and 23 recordings from preoperative patients with a cervical radicular syndrome. Recordings were captured using one of three dynamic X-ray recording systems: Siemens Arcadis Avantic VC10A (Siemens AG, Munich, Germany), Toshiba Infinix Vi (Ōtawara-shi, Tochigi-ken, Japan), or Philips Allura Xper FD20 X-ray system (Best, The Netherlands). All systems recorded at a resolution of 1024 × 1024 pixels and a frame rate of 15 frames per second. No compression was applied to the images. The recordings were split into subsets for training, validation, and testing, with a split of 60-20-20 (Table 1).

2.3. Development of the Model

A fully automated segmentation approach was implemented using nnU-Net for cervical vertebrae detection in dynamic X-ray recordings. nnU-Net is a self-configuring deep learning framework that automatically adapts its architecture, preprocessing, training, and postprocessing to the characteristics of a given dataset without the need for manual tuning [7]. We used the second generation of the nnU-Net model (v2), which offers improved modularity and reproducibility. In this work, the framework was used in default configuration mode, which automatically selects optimal network architecture, patch and batch sizes, and augmentation strategies. The only manual changes were the reduction in the number of training epochs, which refers to one complete pass through the entire training dataset, to 100 and the adoption of the residual encoder architecture. The standard 1000 epochs resulted in overfitting of the model and model convergence was consistently achieved within 100 epochs. The residual encoder improved segmentation performance across all model scales [13].
This architecture integrates ResNet-style blocks into the U-Net encoder to enhance feature extraction and gradient propagation. Residual skip connections allow the network to learn residual mappings, mitigating vanishing gradients and enabling deeper architectures. The nnU-Net ResEnc M variant was selected due to computational constraints.

2.4. Histogram Equalization

To improve anatomical visibility, histogram equalization was applied as a preprocessing step, redistributing pixel intensity values to enhance contrast and highlight structural boundaries. To avoid amplifying background noise or artifacts, binary masks isolating the patient region were generated and equalization was applied only within these regions.
For recordings acquired on the Siemens systems, collimators visible in the image prevented reliable threshold-based masking of the subject. In such cases, histogram matching was performed instead of histogram equalization, aligning each sequence to a reference frame with previously demonstrated high segmentation performance [6]. In total 20 Siemens system recordings were incorporated into the dataset, accounting for 18% of the total data. The recordings were split with the same proportions among the three subsets as the global distribution. Figure 1 illustrates the differences between the image acquisition systems and provides examples of histogram equalization and histogram matching.

2.5. Pre-Training

Given the limited availability of annotated dynamic X-ray data, a synthetic dataset was created from the CSXA (Cervical Spine X-ray Anatomy) dataset [14]. Static cervical spine images were adapted to resemble dynamic sequences by adding noise and matching their histograms to the dynamic dataset. Each adapted image was matched to a randomly selected frame from the training or validation sets.
This synthetic dataset was used to pre-train a separate nnU-Net for 1000 epochs under default settings, except for the random rotation augmentation, which was expanded to rotate all training samples by random angles between –90° and +90°. This modification improves robustness to orientation variability, compensating for the neutral position bias of the CSXA.
The model weights were then finetuned for another 100 epochs on the target dataset. Four nnU-Net configurations were tested: baseline (A), pre-trained (B), histogram equalization (C) and pre-trained with histogram equalization (D).

2.6. Model Evaluation Metrics

Included evaluation metrics were Dice Similarity Coefficient (DSC), Intersection over Union (IoU), and 95th percentile Hausdorff Distance (HD95). The DSC measures the overlap between predicted (A) and ground-truth (B) segmentations, as in Equation (1).
D S C   ( A , B ) = 2 × | A B | | A | + | B |
A value of 1 indicates perfect overlap and 0 indicates no overlap. While scores above 0.7 are generally considered acceptable in medical image segmentation, a stricter threshold of 0.8 was adopted in this study to ensure rotation accuracy [15].
The IoU, also known as the Jaccard Index, is defined as the ratio of the area of the intersection over the area of the union of the segmentation and the ground truth, as in Equation (2).
I o U   ( A , B ) = | A B | | A B |
This measure ranges from 0 to 1 and is more sensitive to segmentation errors than Dice; here, IoU scores above 0.7 were considered sufficient [16].
The HD95 quantifies the maximum distance between the boundaries of the predicted and ground-truth segmentations, using the 95th percentile to reduce the influence of outliers, as in Equation (3)
H D 95 A , B = m a x ( sup i n f ( A , B ) , sup i n f ( B , A ) ) 95
Here, d(A, B) represents the Euclidean distance between boundary points. In this study, HD95 values below 16 mm, approximately the diameter of a vertebral body, were considered acceptable [17].

2.7. Mean Shape

In dynamic X-ray recordings, vertebrae are rigid anatomical structures and should maintain a consistent shape throughout motion. nnU-Net may produce slight variations in vertebral shape across frames due to noise, occlusion, coupled motion, or model uncertainty. To address this, a mean shape was constructed for each vertebra per recording to enforce anatomical consistency and improve reliability. The skull base (C0) was excluded due to its inconsistent visibility and variable segmentation quality across frames.

2.8. Outcome

The relative rotation between adjacent vertebrae was estimated using mean shapes, constructed as previously described [6]. For each frame in a sequence, the mean shape of a vertebra was translated to the centroid of the segmentation predicted by the model and iteratively rotated to achieve maximal overlap. In some cases, the previous iteration of the algorithm would find an optimum at an angle around 180 degrees higher or lower than the previous frame. To further increase the consistency of the algorithm, the range of angles was limited to ±90 degrees, which prevented these flips in subsequent frames. The fitted mean shape was used to calculate the rotation of each vertebra C in the consecutive frames of the entire recording (dθC), as shown in Equation (4), where θ denotes an angle in frame t.
d θ C ( t ) = θ   C ( t ) θ   C ( t 1 )
C(t) represents the change in absolute rotation between consecutive frames. For the first frame, dθC(0) was set to zero. With the relative rotation values of each vertebra, the relative rotation between two vertebrae (dθR(t)) was calculated via Equation (5). With Ck and Cl each indicating another vertebra.
d θ R ( t ) = d θ C k ( t ) d θ C l ( t )
To suppress high-frequency noise in the rotation signal, a low-pass Gaussian filter was applied to the rotation time series. The optimal sigma (σ) value was determined for all vertebral segments and was set to a value of 3, preserving overall motion trends while reducing noise-induced fluctuations.

2.9. Analyses

The Intraclass Correlation Coefficient (ICC) was used to quantify agreement between predicted and manually annotated relative rotation values. A two-way mixed-effects model with consistency agreement was applied [18]. The ICC (3, 1) model was chosen as it assesses the similarity in rotational patterns between raters instead of the absolute mean differences. ICC values above 0.6 were deemed acceptable, and those above 0.8 were considered excellent for clinical use.
Sensitivity analyses were performed to investigate the potential influence of including low-quality recordings and recordings with low sROM. Recording quality was visually assessed by two trained individuals, based on the presence of noise and overprojection. The segments that were classified as low-quality were excluded to use as a sensitivity analysis. To determine whether low sROM had an influence, vertebral segments with a ROM below 4 degrees were excluded as a sensitivity analysis.
To analyze motion patterns, the segmental rotation between each successive frame of individual motion segments from C4 to C7 was plotted against the cumulative rotation of C4 to C7. The resulting graphs were then generated and compared with the ground truth.

2.10. Correlation Analysis

To provide a further indication of the relation between the quality of the segmentation and the ICC of each recording, the Pearson correlation coefficient was calculated between the ICC and the mean, median and standard deviation of each segmentation metric across each recording. With these correlations, the impact of the average quality and the variability of the segmentations on the reliability of the motion pattern analysis can be determined.

3. Results

3.1. Segmentation Performance

The mean DSC ranged from 0.69 to 0.91 for the baseline Model (A), from 0.71 to 0.92 for the pre-trained Model (B), from 0.69 to 0.90 for the histogram equalization Model (C), and from 0.67 to 0.91 for the pre-trained with histogram equalization Model (D). The mean IoU ranged from 0.57 to 0.84 for Model A, from 0.58 to 0.85 for Model B, from 0.57 to 0.83 for Model C, and from 0.55 to 0.84 for Model D. The mean HD95 ranged from 2.49 to 19.55 mm for Model A, from 2.35 to 19.67 mm for Model B, from 3.78 to 19.01 mm for Model C, and from 2.58 to 19.09 mm for Model D. The mean and standard deviation of the DSC, IoU, and HD95 for each vertebra across all models are presented in Table 2, calculated across all frames. Figure 2 shows an example of the segmentation performance, comparing the ground truth with the results produced by Models A–D.

3.2. Intraclass Correlation Coefficient

The ICC values for vertebral segments ranged from 0.449 to 0.728 for Model A, from 0.368 to 0.734 for Model B, from 0.287 to 0.612 for Model C, and from 0.382 to 0.742 for Model D, presented in Table 3. All recordings were included, except one in Model B and one in Model C, where the mean shape calculation failed, precluding the computation of rotations.

3.3. Sensitivity Analyses

After excluding segments due to low sROM, the ICC values ranged from 0.516 to 0.728 for Model A, from 0.399 to 0.763 for Model B, from 0.321 to 0.612 for Model C, and from 0.442 to 0.742 for Model D, presented in Table 4. After excluding segments due to low-quality, the ICC values ranged from 0.549 to 0.770 for Model A, from 0.532 to 0.761 for Model B, from 0.421 to 0.689 for Model C, and from 0.421 to 0.764 for Model D, presented in Table 5. After excluding segments due to low sROM and low-quality, the ICC values ranged from 0.617 to 0.837 for model A, from 0.609 to 0.780 for model B, from 0.409 to 0.689 for model C, and from 0.480 to 0.835 for model D, presented in Table 6. As none of the C3–C4 and C4–C5 segments had an sROM below 4 degrees, they were not filtered out, resulting in identical values in Table 3, Table 4, Table 5 and Table 6.

3.4. Motion Pattern Analysis

The motion pattern analyzed with Model A with the highest mean ICC is shown in Figure 3 and the motion pattern with the lowest mean ICC is shown in Figure 4. All motion patterns analyzed with Model A are presented in Appendix A (Figure A1, Figure A2, Figure A3, Figure A4, Figure A5, Figure A6, Figure A7, Figure A8, Figure A9, Figure A10, Figure A11, Figure A12, Figure A13, Figure A14, Figure A15, Figure A16, Figure A17, Figure A18, Figure A19, Figure A20 and Figure A21).

3.5. ICC vs. Segmentation Performance Correlation

In Table 7, the mean median and standard deviation of each segmentation metric is correlated against the ICC of each recording. None of the correlations are significant due to the low number of recordings in the dataset.

4. Discussion

This study aimed to test and improve nnU-net model performances on cervical dynamic X-ray recordings to support reliable motion analysis in research and clinical settings. Following sensitivity analyses, excluding low sROM and low-quality segments, the ICC values exceeded 0.6 for both the baseline Model (Model A) and the pre-trained Model (Model B), indicating that both models are suitable for motion pattern analysis.
Our findings suggest that pre-training may improve model performance, particularly for lower-quality recordings (Table 5), while the baseline model appears more effective for higher-quality recordings (Table 6 and Table 7). This implies that pre-training could be beneficial for lower-quality data, although its utility is constrained by data scarcity. Napravnik et al. evaluated pre-training across a broad range of datasets for application to more specific tasks and, consistent with our results, reported no significant improvements [19]. Closer alignment of the pre-training dataset with the downstream application may enhance model performance. In this context, pre-training on dynamic rather than static imaging could benefit the model, as the static images can only aid the model in learning vertebral shape, contrast characteristics, and general anatomy Any visual features emerging due to movement are lacking from this data. Future work may also focus on developing refined pre-training strategies that reduce or eliminate the need for human annotation [8].
Histogram equalization appeared to reduce model performance, despite visually enhancing image contrast, as seen in Figure 1. This finding was unexpected, as previous studies have reported histogram equalization to be beneficial [9]. One possible explanation is the overprojection of the shoulders over the lower vertebrae. The resulting large areas of darker pixels may skew the pixel-value histogram and interfere with the equalization process. The resulting large areas of darker pixels may skew the pixel-value histogram and interfere with the equalization process. Histogram equalization was not applied to the Siemens recordings because the visible collimators interfered with masking, yet these recordings corresponded to some of the highest model performances. The presence of collimators likely improved image contrast by focusing the beam on the central area, thereby supporting performance. Additionally, Siemens recordings lacked shoulder overprojection, which in other datasets may have been exacerbated by photon starvation, leading to magnified noise during image reconstruction [20]. Across all models, segmentation performance was generally lower for the skull base (C0) and the lower vertebrae (C6–C7). Although the skull landmark used for the ground truth is visually distinct to human observers, the resulting straight-line annotation lacks a true contrast boundary, making it more difficult for the model to predict.
Despite Model B achieving the highest segmentation performance, with DSC rated as excellent across all levels except C0, overall consistency was limited by variability in manual annotations. The deviating morphology of the C2 vertebra contributed to inconsistent labeling, as the odontoid was included in some cases but omitted in others. Similarly, the posterior landmark of the processus spinosus was annotated inconsistently, alternating between the rounded surface and the bifid end. These inconsistencies introduced ambiguity for the model and contributed to lower ICC values for the C2–C3 segment. Segment C6–C7 also demonstrated lower ICC values. Shoulder overprojection frequently affected visualization of this segment, complicating accurate annotation for both the models but also for human observers. The relationship between segmentation performance and ICC, shown in Table 7, indicates that for most models, improved segmentation quality does not significantly increase the reliability of angle determination. However, model C has higher correlation levels than the other models, while also being the lowest performing segmentation model. This could indicate that poorer segmentation quality does impact the ICC, but with diminishing returns after a certain threshold. Sensitivity analyses demonstrated that model performance depended on recording quality, in particular high contrast and the absence of shoulder overprojection. Figure A22 highlights the contrast between the worst- and best-performing cases. While low contrast can be mitigated by increasing tube voltage, this improvement comes at the cost of higher radiation exposure, but also lower contrast in other parts of the image, for instance midcervical.
Improved ICC values observed after excluding recordings with low sROM may be explained by a lower signal-to-noise ratio in the excluded recordings. When motion amplitudes are small, the true movement between frames is similar in magnitude to imaging noise and annotation variability, causing minor segmentation inaccuracies to disproportionately affect derived relative rotation measures. The sensitivity analysis indicates that low sROM constitutes a methodological limitation for reliable motion analysis.
To mitigate large deviations in segmentation, a promptable model could offer a potential solution. A promptable model requires human input, while models with automatic inference do not. The Segment Anything Model 2 (SAM2) was also evaluated by our study team. SAM2 is a promptable segmentation model designed for both static images and video data [21]. We hypothesized that the use of manual prompts to indicate areas of interest could reduce large deviations in segmentation while substantially lowering the labor intensity of the annotation process. The segmentation and ICC results of the SAM2 model are provided in Appendix A, Table A1 and Table A2. Table A1 shows that the SAM2 segmentation results lie between those of models A and B. However, SAM2 yielded lower ICC values and required longer analysis times than the fully automated nnU-Net models. As the original SAM2 model was not trained on medical data, it may not be suitable for our dataset, even though the model was fine-tuned on dynamic X-ray recordings.
The motion pattern graphs show that, while fine details are not consistently tracked by model A, larger trends are captured well. From a biomechanical perspective, the cervical spine is not expected to exhibit continuous small oscillations during flexion and extension, which suggests that the model may, under certain conditions, provide estimates that are more physiologically plausible than the manual annotations. The lowest performing recording illustrates this limitation, as the model exhibited difficulties in reproducing rapidly changing relative rotations observed in the ground truth. Notably, this recording was also part of the poor image quality subset, which likely contributed to the reduced performance. In this case, the model’s prediction could also be more accurate than the gold standard.
Previous studies have demonstrated that motion patterns can differentiate asymptomatic individuals from patients with cervical radiculopathy [1]. Based on these observations, several clinically meaningful applications can be anticipated, particularly in the evaluation of spinal instability, the longitudinal monitoring of degenerative changes, and the assessment of post-surgical biomechanics. For example, specific motion patterns, rather than only the presence or absence of motion, may help to identify segments with a higher likelihood of heterotopic ossification, which could inform the choice between a motion-preserving implant and fusion implant. In this context, pre- and post-laminectomy analyses could offer valuable insights into how decompression alters segmental mobility and spinal alignment, which are currently assessed primarily using static imaging [22]. Likewise, longitudinal assessments of motion patterns before and after surgery may help to identify early indicators of complications such as adjacent segment disease (ASD) or post laminectomy instability.
The integration of large-scale datasets with qualitative motion analysis also holds promise for uncovering previously unrecognized kinematic patterns that may influence clinical outcomes and refine treatment strategies. Ultimately, these applications underscore the potential of AI-based motion tracking not only as a diagnostic adjunct but also as a decision-support system tailored to the individual patient. By linking qualitative motion analysis to clinical outcomes, this approach may contribute to more personalized, evidence-based management of cervical spine disorders.

5. Conclusions

This study demonstrates that the tested nnU-Net is capable of accurately capturing motion patterns of the cervical spine provided recording quality is sufficient and a range of motion of at least 4 degrees per vertebral segment. For potential future applications in both research and clinical practice, high image contrast and adequate sROM are recommended to ensure robust model performance. However, as these conditions may not consistently be met in routine clinical practice, broader applicability remains to be demonstrated. Moreover, in anatomically altered cases, such as patients with implants or following laminectomy, further model training will be necessary to maintain accuracy. Overall, this study provides an important step toward automated qualitative motion analysis, making the evaluation of motion patterns more accessible and scalable.

Author Contributions

Conceptualization, E.v.S., T.H.W.H. and T.F.M.B.; methodology, E.v.S. and T.H.W.H.; software, V.N.E.S., T.H.W.H. and T.F.M.B.; validation, E.v.S. and T.H.W.H.; formal analysis, T.H.W.H.; investigation, E.v.S. and T.H.W.H.; resources, V.N.E.S. and T.F.M.B.; data curation, T.H.W.H.; writing—original draft preparation, E.v.S.; writing—review and editing, T.H.W.H., V.N.E.S., A.Y.J.M.S., H.v.S., R.d.B., M.V. and T.F.M.B.; visualization, E.v.S. and T.H.W.H.; supervision, V.N.E.S., A.Y.J.M.S., H.v.S., R.d.B., M.V. and T.F.M.B.; project administration, E.v.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

The clinical studies of which the data have been obtained have all been approved by the Medical Ethical Committee of Maastricht University Medical Center (protocol code: NCT00868335, NL14240.068.06/METC-06-1-098, approval date: 2 May 2007), the Ethical Committee of Zuyderland Medical Center (protocol code: NCT04222777, NL70381.096.19/Z2019087, approval date: 10 October 2019) and National Research Ethics Service Committee South West—Cornwall & Plymouth (protocol code: 11/SW/0072, approval date: 3 May 2011). The studies were conducted according to principles enshrined in the Declaration of Helsinki and in accordance with the Medical Research Involving Human Subjects Act (WMO).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The data presented in this study are available on request from the corresponding author due to privacy restrictions.

Acknowledgments

We would like to thank the subjects who participated in the studies.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
ACDAnterior Cervical Discectomy
ACDAAnterior Cervical Discectomy with Arthroplasty
AIArtificial intelligence 
DSCDice similarity coefficient
HD95Hausdorff Distance
ICCIntraclass Correlation Coefficient
IoUIntersection over Union
sROMSegmental range of motion
SSCSequence of segmental contribution

Appendix A

Figure A1. Motion pattern of HC02 T1e (Model A, ICC = 0.759).
Figure A1. Motion pattern of HC02 T1e (Model A, ICC = 0.759).
Bioengineering 13 00351 g0a1
Figure A2. Motion pattern of HC02 T2e (Model A, ICC = 0.571).
Figure A2. Motion pattern of HC02 T2e (Model A, ICC = 0.571).
Bioengineering 13 00351 g0a2
Figure A3. Motion pattern of HC03 T1e (Model A, ICC = 0.865).
Figure A3. Motion pattern of HC03 T1e (Model A, ICC = 0.865).
Bioengineering 13 00351 g0a3
Figure A4. Motion pattern of HC04 T2e (Model A, ICC = 0.514).
Figure A4. Motion pattern of HC04 T2e (Model A, ICC = 0.514).
Bioengineering 13 00351 g0a4
Figure A5. Motion pattern of HC13 T1e (Model A, ICC = 0.790).
Figure A5. Motion pattern of HC13 T1e (Model A, ICC = 0.790).
Bioengineering 13 00351 g0a5
Figure A6. Motion pattern of HC13 T1f (Model A, ICC = 0.841).
Figure A6. Motion pattern of HC13 T1f (Model A, ICC = 0.841).
Bioengineering 13 00351 g0a6
Figure A7. Motion pattern of HC13 T2f (Model A, ICC = 0.688).
Figure A7. Motion pattern of HC13 T2f (Model A, ICC = 0.688).
Bioengineering 13 00351 g0a7
Figure A8. Motion pattern of HC14 T1e (Model A, ICC = 0.839).
Figure A8. Motion pattern of HC14 T1e (Model A, ICC = 0.839).
Bioengineering 13 00351 g0a8
Figure A9. Motion pattern of HC14 T1f (Model A, ICC = 0.859).
Figure A9. Motion pattern of HC14 T1f (Model A, ICC = 0.859).
Bioengineering 13 00351 g0a9
Figure A10. Motion pattern of HC14 T2e (Model A, ICC = 0.804).
Figure A10. Motion pattern of HC14 T2e (Model A, ICC = 0.804).
Bioengineering 13 00351 g0a10
Figure A11. Motion pattern of HC14 T2f (Model A, ICC = 0.779).
Figure A11. Motion pattern of HC14 T2f (Model A, ICC = 0.779).
Bioengineering 13 00351 g0a11
Figure A12. Motion pattern of HC25 T2e (Model A, ICC = 0.369).
Figure A12. Motion pattern of HC25 T2e (Model A, ICC = 0.369).
Bioengineering 13 00351 g0a12
Figure A13. Motion pattern of HV12 T1e (Model A, ICC = 0.409).
Figure A13. Motion pattern of HV12 T1e (Model A, ICC = 0.409).
Bioengineering 13 00351 g0a13
Figure A14. Motion pattern of HV12 T2e (Model A, ICC = 0.214).
Figure A14. Motion pattern of HV12 T2e (Model A, ICC = 0.214).
Bioengineering 13 00351 g0a14
Figure A15. Motion pattern of HV22 T1e (Model A, ICC = 0.210).
Figure A15. Motion pattern of HV22 T1e (Model A, ICC = 0.210).
Bioengineering 13 00351 g0a15
Figure A16. Motion pattern of HV22 T2e (Model A, ICC = 0.406).
Figure A16. Motion pattern of HV22 T2e (Model A, ICC = 0.406).
Bioengineering 13 00351 g0a16
Figure A17. Motion pattern of HV24 T1e (Model A, ICC = 0.656).
Figure A17. Motion pattern of HV24 T1e (Model A, ICC = 0.656).
Bioengineering 13 00351 g0a17
Figure A18. Motion pattern of P02 T1e (Model A, ICC = 0.875).
Figure A18. Motion pattern of P02 T1e (Model A, ICC = 0.875).
Bioengineering 13 00351 g0a18
Figure A19. Motion pattern of P08 T1e (Model A, ICC = 0.159).
Figure A19. Motion pattern of P08 T1e (Model A, ICC = 0.159).
Bioengineering 13 00351 g0a19
Figure A20. Motion pattern of P14 T1e (Model A, ICC = 0.776).
Figure A20. Motion pattern of P14 T1e (Model A, ICC = 0.776).
Bioengineering 13 00351 g0a20
Figure A21. Motion pattern of P15 T1e (Model A, ICC = 0.506).
Figure A21. Motion pattern of P15 T1e (Model A, ICC = 0.506).
Bioengineering 13 00351 g0a21
Figure A22. Dynamic X-ray images of the worst and best performing recordings in terms of ICC (Left: HC30 T1e, ICC 0.038. Right: HC03 T2e, ICC 0.918).
Figure A22. Dynamic X-ray images of the worst and best performing recordings in terms of ICC (Left: HC30 T1e, ICC 0.038. Right: HC03 T2e, ICC 0.918).
Bioengineering 13 00351 g0a22
Table A1. Segmentation performance of the SAM2 model on the test set.
Table A1. Segmentation performance of the SAM2 model on the test set.
DSC (SD)IoU (SD)HD95 in mm (SD)
C00.69 ± 0.250.57 ± 0.2519.55 ± 16.06
C10.86 ± 0.080.76 ± 0.114.19 ± 2.48
C20.88 ± 0.060.79 ± 0.094.47 ± 2.04
C30.90 ± 0.050.82 ± 0.073.19 ± 2.42
C40.91 ± 0.050.84 ± 0.062.49 ± 1.71
C50.88 ± 0.070.78 ± 0.104.12 ± 3.67
C60.82 ± 0.150.72 ± 0.186.82 ± 5.92
C70.75 ± 0.190.63 ± 0.2110.06 ± 8.71
IoU = Intersection over Union, DSC = Dice Similarity Coefficient, 95th percentile Hausdorff Distance (HD95), SD = standard deviation.
Table A2. Sensitivity analysis for the SAM2 model. ICC values for vertebral segments calculated using the mean shape algorithm.
Table A2. Sensitivity analysis for the SAM2 model. ICC values for vertebral segments calculated using the mean shape algorithm.
ExclusionNonenLow-QualitynLow sROMnLow-Quality/sROMn
Segment
C1–C20.589
(−0.074–0.949)
230.625
(−0.074–0.949)
190.681
(0.165–0.949)
190.750
(0.491–0.949)
15
C2–C30.423
(−0.444–0.941)
230.491
(−0.444–0.941)
190.469
(−0.444–0.941)
200.539
(−0.444–0.941)
17
C3–C40.679
(0.031–0.948)
230.728
(0.158–0.948)
190.679
(0.031–0.948)
230.728
(0.158–0.948)
19
C4–C50.681
(0.123–0.963)
230.679
(0.123–0.963)
170.681
(0.123–0.963)
230.679
(0.123–0.963)
17
C5–C60.676
(−0.003–0.982)
230.695
(−0.003–0.982)
160.672
(−0.003–0.982)
220.689
(−0.003–0.982)
15
C6–C70.385
(−0.431–0.989)
230.522
(−0.095–0.989)
90.391
(−0.431–0.989)
210.556
(−0.095–0.989)
8
ICC = Intraclass Correlation Coefficient.

References

  1. Boselie, T.F.M.; van Santbrink, H.; de Bie, R.A.; van Mameren, H. Pilot Study of Sequence of Segmental Contributions in the Lower Cervical Spine During Active Extension and Flexion: Healthy Controls Versus Cervical Degenerative Disc Disease Patients. Spine 2017, 42, E642–E647. [Google Scholar] [CrossRef] [PubMed]
  2. Van Mameren, H.; Drukker, J.; Sanches, H.; Beursgens, J. Cervical spine motion in the sagittal plane (I) range of motion of actually performed movements, an X-ray cinematographic study. Eur. J. Morphol. 1990, 28, 47–68. [Google Scholar] [PubMed]
  3. Boselie, T.F.; van Mameren, H.; de Bie, R.A.; van Santbrink, H. Cervical spine kinematics after anterior cervical discectomy with or without implantation of a mobile cervical disc prosthesis; an RCT. BMC Musculoskelet. Disord. 2015, 16, 34. [Google Scholar] [CrossRef] [PubMed]
  4. Schuermans, V.N.E.; Smeets, A.; Curfs, I.; van Santbrink, H.; Boselie, T.F.M. A randomized controlled trial with extended long-term follow-up: Quality of cervical spine motion after anterior cervical discectomy (ACD) or anterior cervical discectomy with arthroplasty (ACDA). Brain Spine 2024, 4, 102726. [Google Scholar] [CrossRef] [PubMed]
  5. Schuermans, V.N.E.; Smeets, A.; Breen, A.; Branney, J.; Curfs, I.; van Santbrink, H.; Boselie, T.F.M. An observational study of quality of motion in the aging cervical spine: Sequence of segmental contributions in dynamic fluoroscopy recordings. BMC Musculoskelet. Disord. 2024, 25, 330. [Google Scholar] [CrossRef] [PubMed]
  6. van Santbrink, E.; Schuermans, V.; Cerfonteijn, E.; Breeuwer, M.; Smeets, A.; van Santbrink, H.; Boselie, T. AI-Assisted Image Recognition of Cervical Spine Vertebrae in Dynamic X-Ray Recordings. Bioengineering 2025, 12, 679. [Google Scholar] [CrossRef] [PubMed]
  7. Isensee, F.; Jaeger, P.F.; Kohl, S.A.A.; Petersen, J.; Maier-Hein, K.H. nnU-Net: A self-configuring method for deep learning-based biomedical image segmentation. Nat. Methods 2021, 18, 203–211. [Google Scholar] [CrossRef] [PubMed]
  8. Han, X.; Zhang, Z.; Ding, N.; Gu, Y.; Liu, X.; Huo, Y.; Qiu, J.; Yao, Y.; Zhang, A.; Zhang, L.; et al. Pre-trained models: Past, present and future. AI Open 2021, 2, 225–250. [Google Scholar] [CrossRef]
  9. Saifullah, S.; Dreżewski, R. Modified Histogram Equalization for Improved CNN Medical Image Segmentation. Procedia Comput. Sci. 2023, 225, 3021–3030. [Google Scholar] [CrossRef]
  10. Branney, J. An Observational Study of Changes in Cervical Inter-Vertebral Motion and the Relationship with Patient-Reported Outcomes in Patients Undergoing Spinal Manipulative Therapy for Neck Pain. Ph.D. Thesis, Bournemouth University, Poole, UK, 2014. [Google Scholar]
  11. Branney, J.; Breen, A.C. Does inter-vertebral range of motion increase after spinal manipulation? A prospective cohort study. Chiropr. Man. Ther. 2014, 22, 24. [Google Scholar] [CrossRef] [PubMed]
  12. Reinartz, R.; Platel, B.; Boselie, T.; van Mameren, H.; van Santbrink, H.; Romeny, B. Cervical vertebrae tracking in video-fluoroscopy using the normalized gradient field. In Medical Image Computing and Computer-Assisted Intervention; Springer: Berlin/Heidelberg, Germany, 2009; pp. 524–531. [Google Scholar] [CrossRef]
  13. Isensee, F.; Wald, T.; Ulrich, C.; Baumgartner, M.; Roy, S.; Maier-Hein, K.; Jaeger, P.F. nnU-Net Revisited: A Call for Rigorous Validation in 3D Medical Image Segmentation. arXiv 2024, arXiv:2404.09556. [Google Scholar] [CrossRef]
  14. Ran, Y.; Qin, W.; Qin, C.; Li, X.; Liu, Y.; Xu, L.; Mu, X.; Yan, L.; Wang, B.; Dai, Y.; et al. A high-quality dataset featuring classified and annotated cervical spine X-ray atlas. Sci. Data 2024, 11, 625. [Google Scholar] [CrossRef] [PubMed]
  15. Boehringer, A.S.; Sanaat, A.; Arabi, H.; Zaidi, H. An active learning approach to train a deep learning algorithm for tumor segmentation from brain MR images. Insights Imaging 2023, 14, 141. [Google Scholar] [CrossRef] [PubMed]
  16. Gon Park, S.; Park, J.; Rock Choi, H.; Ho Lee, J.; Tae Cho, S.; Goo Lee, Y.; Ahn, H.; Pak, S. Deep Learning Model for Real-time Semantic Segmentation During Intraoperative Robotic Prostatectomy. Eur. Urol. Open Sci. 2024, 62, 47–53. [Google Scholar] [CrossRef] [PubMed]
  17. Ramieri, A.; Costanzo, G.; Miscusi, M. Functional Anatomy and Biomechanics of the Cervical Spine. In Cervical Spine: Minimally Invasive and Open Surgery; Menchetti, P.P.M., Ed.; Springer International Publishing: Cham, Switzerland, 2022; pp. 11–31. [Google Scholar] [CrossRef]
  18. Koo, T.K.; Li, M.Y. A Guideline of Selecting and Reporting Intraclass Correlation Coefficients for Reliability Research. J. Chiropr. Med. 2016, 15, 155–163. [Google Scholar] [CrossRef] [PubMed]
  19. Napravnik, M.; Hržić, F.; Urschler, M.; Miletić, D.; Štajduhar, I. Lessons learned from RadiologyNET foundation models for transfer learning in medical radiology. Sci. Rep. 2025, 15, 21622. [Google Scholar] [CrossRef] [PubMed]
  20. Barrett, J.F.; Keat, N. Artifacts in CT: Recognition and Avoidance. RadioGraphics 2004, 24, 1679–1691. [Google Scholar] [CrossRef] [PubMed]
  21. Ravi, N.; Gabeur, V.; Hu, Y.-T.; Hu, R.; Ryali, C.; Ma, T.; Khedr, H.; Rädle, R.; Rolland, C.; Gustafson, L.; et al. SAM 2: Segment Anything in Images and Videos. arXiv 2024, arXiv:2408.00714. [Google Scholar] [CrossRef] [PubMed]
  22. Machino, M.; Yukawa, Y.; Hida, T.; Ito, K.; Nakashima, H.; Kanbara, S.; Morita, D.; Kato, F. Cervical alignment and range of motion after laminoplasty: Radiographical data from more than 500 cases with cervical spondylotic myelopathy and a review of the literature. Spine 2012, 37, E1243–E1250. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Example of histogram equalization and histogram matching. (a) Pre-processed image before histogram equalization; (b) image after histogram equalization; (c) pre-processed image before histogram matching; (d) image after histogram matching.
Figure 1. Example of histogram equalization and histogram matching. (a) Pre-processed image before histogram equalization; (b) image after histogram equalization; (c) pre-processed image before histogram matching; (d) image after histogram matching.
Bioengineering 13 00351 g001aBioengineering 13 00351 g001b
Figure 2. Example of segmentation performance HC02 T1e; (a) ground truth, (b) model A, (c) model B, (d) model C, (e) model D. The average DSC and ICC across all vertebrae of each model are displayed above every segmentation mask.
Figure 2. Example of segmentation performance HC02 T1e; (a) ground truth, (b) model A, (c) model B, (d) model C, (e) model D. The average DSC and ICC across all vertebrae of each model are displayed above every segmentation mask.
Bioengineering 13 00351 g002
Figure 3. Motion pattern with the highest mean ICC (Model A, HC03 T2e ICC = 0.918). The left figure illustrates the ground truth motion pattern, whereas the right figure illustrates the motion pattern computed by the model. Extension motion over time is plotted on the x-axis, and relative rotations are plotted on the y-axis.
Figure 3. Motion pattern with the highest mean ICC (Model A, HC03 T2e ICC = 0.918). The left figure illustrates the ground truth motion pattern, whereas the right figure illustrates the motion pattern computed by the model. Extension motion over time is plotted on the x-axis, and relative rotations are plotted on the y-axis.
Bioengineering 13 00351 g003
Figure 4. Motion pattern with the lowest mean ICC (Model A, HC30 T1e ICC = 0.038). The left figure illustrates the ground truth motion pattern, whereas the right figure illustrates the motion pattern computed by the model. Extension motion over time is plotted on the x-axis, and relative rotations are plotted on the y-axis.
Figure 4. Motion pattern with the lowest mean ICC (Model A, HC30 T1e ICC = 0.038). The left figure illustrates the ground truth motion pattern, whereas the right figure illustrates the motion pattern computed by the model. Extension motion over time is plotted on the x-axis, and relative rotations are plotted on the y-axis.
Bioengineering 13 00351 g004
Table 1. Overview of the included data divided into subsets, displayed for number of individuals and number of recordings.
Table 1. Overview of the included data divided into subsets, displayed for number of individuals and number of recordings.
Data SubsetIndividuals (N)Recordings (N)
Training (60%)3667
Validation (20%)1222
Testing (20%)1423
Table 2. Segmentation performance of model options A to D on the test set.
Table 2. Segmentation performance of model options A to D on the test set.
DSC (SD)IoU (SD)HD95 in mm (SD)
ModelABCDABCDABCD
C00.69 ± 0.250.71 ± 0.220.69 ± 0.240.68 ± 0.230.57 ± 0.250.58 ± 0.230.57 ± 0.250.55 ± 0.2319.55 ± 16.0619.67 ± 18.2019.07 ± 17.9419.25 ± 18.70
C10.86 ± 0.080.86 ± 0.080.85 ± 0.100.87 ± 0.040.76 ± 0.110.76 ± 0.100.74 ± 0.110.78 ± 0.064.19 ± 2.484.07 ± 1.554.74 ± 5.563.93 ± 1.50
C20.88 ± 0.060.89 ± 0.040.87 ± 0.080.89 ± 0.040.79 ± 0.090.80 ± 0.070.78 ± 0.100.80 ± 0.064.47 ± 2.044.15 ± 1.736.01 ± 10.524.31 ± 1.81
C30.90 ± 0.050.91 ± 0.030.90 ± 0.060.90 ± 0.030.82 ± 0.070.83 ± 0.050.82 ± 0.080.82 ± 0.053.19 ± 2.422.76 ± 1.183.80 ± 8.352.99 ± 1.68
C40.91 ± 0.050.92 ± 0.030.90 ± 0.060.91 ± 0.030.84 ± 0.060.85 ± 0.040.83 ± 0.080.84 ± 0.052.49 ± 1.712.35 ± 1.393.78 ± 5.852.58 ± 1.61
C50.88 ± 0.070.90 ± 0.040.87 ± 0.080.89 ± 0.040.78 ± 0.100.82 ± 0.060.77 ± 0.110.81 ± 0.064.12 ± 3.673.19 ± 1.835.30 ± 4.783.29 ± 1.67
C60.82 ± 0.150.90 ± 0.040.80 ± 0.170.89 ± 0.050.72 ± 0.180.82 ± 0.070.69 ± 0.190.80 ± 0.076.82 ± 5.923.68 ± 3.197.42 ± 6.354.42 ± 3.35
C70.75 ± 0.190.84 ± 0.100.73 ± 0.210.81 ± 0.130.63 ± 0.210.73 ± 0.140.62 ± 0.220.70 ± 0.1610.06 ± 8.715.55 ± 4.3511.39 ± 15.797.81 ± 12.39
IoU = Intersection over Union, DSC = Dice Similarity Coefficient, 95th percentile Hausdorff Distance (HD95), SD = standard deviation.
Table 3. ICC scores individual vertebrae calculated per model using the mean shape.
Table 3. ICC scores individual vertebrae calculated per model using the mean shape.
ModelAnBnCnDn
Segment
C1–C20.595
(−0.369–0.981)
230.569
(−0.423–0.966)
220.476
(−0.271–0.966)
220.577
(−0.222–0.967)
23
C2–C30.449
(−0.262–0.913)
230.546
(−0.275–0.956)
220.384
(−0.421–0.858)
220.438
(−0.386–0.920)
23
C3–C40.697
(−0.212–0.971)
230.734
(−0.041–0.967)
220.612
(−0.409–0.928)
220.724
(−0.055–0.967)
23
C4–C50.728
(0.086–0.961)
230.715
(0.080–0.965)
220.589
(−0.173–0.955)
220.742
(0.301–0.952)
23
C5–C60.693
(−0.042–0.968)
230.733
(−0.198–0.963)
220.342
(−0.410–0.971)
220.689
(−0.087–0.942)
23
C6–C70.449
(−0.480–0.991)
230.368
(−0.161–0.990)
220.287
(−0.151–0.990)
220.382
(−0.435–0.990)
23
ICC = Intraclass Correlation Coefficient.
Table 4. Sensitivity analysis with low sROM segments excluded. ICC values for vertebral segments calculated per model using the mean shape.
Table 4. Sensitivity analysis with low sROM segments excluded. ICC values for vertebral segments calculated per model using the mean shape.
ModelAnBnCnDn
Segment
C1–C20.711
(−0.228–0.981)
190.706
(−0.423–0.966)
180.557
(−0.027–0.950)
180.720
(−0.208–0.967)
19
C2–C30.532
(−0.262–0.913)
200.583
(−0.275–0.956)
190.455
(−0.421–0.858)
190.488
(−0.386–0.920)
20
C3–C40.697
(−0.212–0.971)
230.734
(−0.041–0.967)
220.612
(−0.409–0.928)
220.724
(−0.055–0.967)
23
C4–C50.728
(0.086–0.961)
230.715
(0.080–0.965)
220.589
(−0.173–0.955)
220.742
(0.301–0.952)
23
C5–C60.690
(−0.042–0.968)
220.763
(−0.198–0.963)
210.330
(−0.410–0.971)
210.680
(−0.087–0.942)
22
C6–C70.516
(−0.387–0.991)
210.399
(−0.161–0.990)
200.321
(−0.151–0.968)
200.442
(−0.367–0.990)
21
ICC = Intraclass Correlation Coefficient.
Table 5. Sensitivity analysis with low-quality segments excluded. ICC values for vertebral segments calculated per model using the mean shape.
Table 5. Sensitivity analysis with low-quality segments excluded. ICC values for vertebral segments calculated per model using the mean shape.
ModelAnBnCnDn
Segment
C1–C20.670
(−0.369–0.981)
190.605
(−0.375–0.966)
190.504
(−0.271–0.950)
180.638
(−0.222–0.967)
19
C2–C30.549
(−0.148–0.913)
190.597
(0.162–0.956)
190.498
(−0.082–0.858)
180.543
(−0.109–0.920)
19
C3–C40.770
(0.302–0.971)
190.761
(0.251–0.967)
190.689
(0.077–0.928)
180.764
(0.304–0.967)
19
C4–C50.750
(0.180–0.961)
170.727
(0.080–0.965)
170.652
(0.091–0.955)
160.746
(0.301–0.952)
17
C5–C60.715
(−0.042–0.968)
160.738
(−0.198–0.963)
160.421
(−0.318–0.971)
160.694
(−0.087–0.937)
16
C6–C70.594
(−0.043–0.991)
90.532
(−0.083–0.990)
90.427
(−0.112–0.968)
90.421
(−0.367–0.990)
9
ICC = Intraclass Correlation Coefficient.
Table 6. Sensitivity analysis with low sROM and low-quality segments excluded. ICC values for vertebral segments calculated per model using the mean shape.
Table 6. Sensitivity analysis with low sROM and low-quality segments excluded. ICC values for vertebral segments calculated per model using the mean shape.
ModelAnBnCnDn
Segment
C1–C20.837
(0.604–0.981)
150.780
(0.512–0.966)
150.617
(−0.027–0.950)
140.835
(0.687–0.967)
15
C2–C30.617
(−0.148–0.913)
170.636
(0.208–0.956)
170.561
(0.050–0.858)
160.594
(−0.109–0.920)
17
C3–C40.770
(0.302–0.971)
190.761
(0.251–0.967)
190.689
(0.077–0.928)
180.764
(0.304–0.967)
19
C4–C50.750
(0.180–0.961)
170.727
(0.080–0.965)
170.652
(0.091–0.955)
160.746
(0.301–0.952)
17
C5–C60.711
(−0.042–0.968)
150.743
(−0.198–0.963)
150.409
(−0.318–0.971)
150.680
(−0.087–0.937)
15
C6–C70.673
(0.028–0.991)
80.609
(−0.066–0.990)
80.493
(−0.112–0.968)
80.480
(−0.367–0.990)
8
ICC = Intraclass Correlation Coefficient.
Table 7. The Pearson correlation coefficient indicating the relation between segmentation performance against ICC.
Table 7. The Pearson correlation coefficient indicating the relation between segmentation performance against ICC.
ModelABCD
Metric n n n n
Dice mean 0.047230.010220.654220.26623
Dice median0.03123−0.081220.529220.02823
Dice std−0.02723−0.04122−0.68322−0.28223
IoU mean0.05423−0.008220.649220.25323
IoU median0.02623−0.096220.528220.02523
IoU std−0.00123−0.00422−0.67422−0.29023
HD95 mean−0.13223−0.20922−0.71022−0.36723
HD95 median−0.105230.01122−0.60022−0.09123
HD95 std−0.12823−0.30822−0.51022−0.21523
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

van Santbrink, E.; Hijzelaar, T.H.W.; Schuermans, V.N.E.; Smeets, A.Y.J.M.; van Santbrink, H.; de Bie, R.; Veta, M.; Boselie, T.F.M. Improved AI-Assisted Image Recognition of Cervical Spine Vertebrae Enables Motion Pattern Analysis in Dynamic X-Ray Recordings. Bioengineering 2026, 13, 351. https://doi.org/10.3390/bioengineering13030351

AMA Style

van Santbrink E, Hijzelaar THW, Schuermans VNE, Smeets AYJM, van Santbrink H, de Bie R, Veta M, Boselie TFM. Improved AI-Assisted Image Recognition of Cervical Spine Vertebrae Enables Motion Pattern Analysis in Dynamic X-Ray Recordings. Bioengineering. 2026; 13(3):351. https://doi.org/10.3390/bioengineering13030351

Chicago/Turabian Style

van Santbrink, Esther, Tijmen H. W. Hijzelaar, Valérie N. E. Schuermans, Anouk Y. J. M. Smeets, Henk van Santbrink, Rob de Bie, Mitko Veta, and Toon F. M. Boselie. 2026. "Improved AI-Assisted Image Recognition of Cervical Spine Vertebrae Enables Motion Pattern Analysis in Dynamic X-Ray Recordings" Bioengineering 13, no. 3: 351. https://doi.org/10.3390/bioengineering13030351

APA Style

van Santbrink, E., Hijzelaar, T. H. W., Schuermans, V. N. E., Smeets, A. Y. J. M., van Santbrink, H., de Bie, R., Veta, M., & Boselie, T. F. M. (2026). Improved AI-Assisted Image Recognition of Cervical Spine Vertebrae Enables Motion Pattern Analysis in Dynamic X-Ray Recordings. Bioengineering, 13(3), 351. https://doi.org/10.3390/bioengineering13030351

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop