Next Article in Journal
Multi-Site Photoplethysmography Technology for Blood Pressure Assessment: Challenges and Recommendations
Next Article in Special Issue
Computed and Subjective Blue Scleral Color Analysis as a Diagnostic Tool for Iron Deficiency: A Pilot Study
Previous Article in Journal
Impaired Glucose Metabolism in People with Extremely Elevated High-Density Lipoprotein Cholesterol and Low Alcohol Consumption: Results of the Kanagawa Investigation of Total Checkup Data from the National Database-3 (KITCHEN-3)
Previous Article in Special Issue
Artificial Intelligence Algorithms and Natural Language Processing for the Recognition of Syncope Patients on Emergency Department Medical Records
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Artificial Intelligence for Automatic Measurement of Sagittal Vertical Axis Using ResUNet Framework

1
aetherAI Co., Ltd., Taipei 115, Taiwan
2
Department of Orthopaedic Surgery, Spine Division, Bone and Joint Research Center, Chang Gung Memorial Hospital and Chang Gung University College of Medicine, Taoyuan 333, Taiwan
3
Department of Medical Imaging and Intervention, Chang Gung Memorial Hospital and Chang Gung University College of Medicine, Taoyuan 333, Taiwan
*
Authors to whom correspondence should be addressed.
The authors contributed equally to this work.
J. Clin. Med. 2019, 8(11), 1826; https://doi.org/10.3390/jcm8111826
Submission received: 6 September 2019 / Revised: 24 October 2019 / Accepted: 25 October 2019 / Published: 1 November 2019
(This article belongs to the Special Issue The Future of Artificial Intelligence in Clinical Medicine)

Abstract

:
We present an automated method for measuring the sagittal vertical axis (SVA) from lateral radiography of whole spine using a convolutional neural network for keypoint detection (ResUNet) with our improved localization method. The algorithm is robust to various clinical conditions, such as degenerative changes or deformities. The ResUNet was trained and evaluated on 990 standing lateral radiographs taken at Chang Gung Memorial Hospital, Linkou and performs SVA measurement with median absolute error of 1.183 ± 0.166 mm. The 5-mm detection rate of the C7 body and the sacrum are 91% and 87%, respectively. The SVA calculation takes approximately 0.2 s per image. The intra-class correlation coefficient of the SVA estimates between the algorithm and physicians of different years of experience ranges from 0.946 to 0.993, indicating an excellent consistency. The superior performance of the proposed method and its high consistency with physicians proved its usefulness for automatic measurement of SVA in clinical settings.

1. Introduction

Adult spinal deformity (ASD) is a broad diagnosis referring to stable asymptomatic curves and disabling deformities in the spine that contribute to pain, weakness, and low health-related quality of life (HRQOL). Although ASD is quite common, the variation and unique pattern of each spinal deformity make reproducible measurement difficult. Previous studies investigating the correlation between radiographic appearances and clinical symptoms yielded rather low predictive power due to highly variable health status [1]. Little correlation between radiographic assessment and questionnaire scores was found for adolescent idiopathic scoliosis [2]. However, in 2005, Glassman et al. showed that sagittal plane balance is the most reliable predictor of clinical symptoms in adults with spinal deformity [3]. Even mild positive sagittal balance results in destructive spinal deformity, clinical symptoms of which deteriorate linearly [4]. Therefore, a surgical plan for restoring sagittal balance is crucial in all spinal reconstructive surgeries [3].
Parameters to describe the sagittal alignment of the spine include the sagittal vertical axis (SVA), thoracic kyphosis, lumbar lordosis, pelvic incidence, pelvic tilt, and sacral slope. SVA, which is the most commonly used measure of the sagittal plane, is obtained from standing lateral radiographs and is defined as the horizontal distance between two landmarks: the center of C7 vertebral body and the posterior superior aspect of the S1 vertebral body. According to Schwab’s classification based on Jean Dubousset‘s cone of economy, the realignment of the spine should aim to meet the criteria of SVA < 50 mm to alleviate sensations of falling over [5]. Consequently, formulating a patient-specific surgical plan requires SVA measurement both preoperatively and postoperatively, using whole-spine lateral radiographs. However, manual SVA estimation on radiographs is rather inconvenient due to lack of easy-to-use tools. To solve this issue, we propose using deep-learning models for fully automatic estimation of SVA.
A well-known family of deep-learning models is the convolutional neural networks (CNNs). CNNs have drawn considerable attention since 2012, as they were found to outperform traditional image processing algorithms on image classification tasks [6]. Since then, CNNs have been increasingly used for medical image analysis [7,8]. A CNN is primarily made of stacked convolution layers, which can be regarded as a series of learnable feature extractors designed for acquiring low-to-high-level features of an image. In general, basic image features, such as blob of colors or edges of different orientations, are learnable by the shallow convolution layers. On the other hand, complex image features, such as appearances of objects, are learnable by the deep convolution layers [9]. Therefore, well-trained convolution layers can be used to extract informative features that are useful for landmark localization or other specific tasks [10,11].
Automatic SVA estimation can generally be regarded as the task of landmark localization, as SVA was defined as the horizontal difference between two anatomical landmarks. In this study, we investigated the performance of ResUNet [12] for automatic vertebrae localization on radiographs. ResUNet is a variant of CNN. Its UNet-like structure [13] combines the contextual information produced by deep layers and better location information produced by shallow layers, allowing better utilization of low- and high-level features. In addition, the encapsulated residual blocks [14] of ResUNet enable better flow of information and can avoid performance deterioration as the network goes deeper. ResUNet has been used for cell segmentation from biomedical images [12] and road extraction from satellite images [15]. To the best of our knowledge, our study is the first to utilize ResUNet for anatomical landmark localization.
Some previous works use CNN for vertebrae segmentation or localization. For example, for 3D CT and MR images, Lessmann et al. [16] used a 3D UNet-like model for vertebrae segmentation. In addition, Wimmer et al. [17] used 3D CNN for vertebrae localization. There were also studies on radiographs: for biplanar radiographs, including A-P and lateral view, Gallbusera et al. [18] used a database collected using the EOS™ imaging system [19] and trained CNN models for each of the landmarks. For lateral spine radiographs, Al Arif et al. [20] applied a UNet model for the localization of cervical vertebral centers.
In this study, we applied ResUNet on plain lateral spine radiographs. Although our algorithm is similar to the one in [20], there are some major differences. First, we did not split images into patches, as patching leads to extra pre- and post-processing steps, and the correlation between landmarks of different patches will be ignored completely. However, if patching was not performed, the large image size would lead to insufficient amount of RAM on GPU. To alleviate this issue, we used a small batch size (batch size = 2) and replaced the widely used batch normalization with the group normalization [21] in our network. Group normalization is known to perform well even when the batch size is small, as it does not perform normalization along the batch axis. Secondly, we let probability maps of landmarks to be predicted separately. Thus, predictions were not on the same map and further post-processing steps involving separation and identification of the landmarks can be eliminated.
Our main contribution in this paper is three-fold. Firstly, we validated that automatic SVA estimation on lateral spine radiography is feasible using the current deep-learning techniques. The procedure of 10-fold cross-validation was applied to validate the effectiveness of the constructed SVA estimator (ResUNet). Secondly, modern networks often take the Gaussian heatmaps centering at landmark locations as the targets of regression [22,23,24]. In this study, we experimented also the use of exponential heatmaps for model selection. Our experiments suggested that, when the heatmap function was narrow, the exponential heatmaps can be more easily learned, as compared with the Gaussian heatmaps. Thirdly, we compared the estimates between ResUNet and human experts of different seniority. The results show that they were in excellent agreement.

2. Materials and Methods

2.1. Data Preparation

We collected 990 whole spine lateral radiographs from 990 patients who had degenerative changes or deformities in the spine from January 2012 to September 2018 at Chang Gung Memorial Hospital, Linkou, Taiwan. The collected radiographs were from different X-ray scanners and the mean pixel resolution was 3552 × 1434 . Poor-quality images such as those with poor image contrast and positioning errors were excluded. Among the collected radiographs, 13% (128) of the images are of children (age < 12), 42% (414) of the images are of young adults (age 13–18) and 45% (448) of the images are of adults (age > 18). Overall, 77% (765) of the images do not have an implant, while 23% (225) do. In the group without implants (765), 56% (428) of the patients have scoliosis, while 44% (337) of patients have degenerative change. In the group with implants (229), 76% (170) of the implants extend from thoracic to lumbar, 21% (48) of the implants extend from lumbar to sacral, 2% (5) of the implants only involve lumbar, and 1% (2) of the implants only involve thoracic. Pediatric and young adult patients diagnosed of scoliosis were enrolled in this study. In terms of sacral morphology, this cohort displayed various degrees of physiological sacral fusions, as well as pathology resulting from scoliosis. The images were anonymized, and standard research protocols were followed. This study was approved by the Institutional Ethical Committee (IRB number: 201801651B0 (1809270049)).
The whole dataset (990 images) was annotated by one senior resident orthopedic surgeon. This surgeon has more than five years of experience for manual measurement of spine parameters. When a difficult case was encountered, the surgeon discussed with an experienced spine surgeon to reach a consensus. This annotated dataset was used for model training and evaluation. An annotated radiograph can be seen in Figure 1. We used a custom written MATLAB GUI program for annotation.
In addition, an experienced radiologist and a junior resident physiciane also provided their annotations for a subset of the dataset (99 images), which were randomly selected from the 990 images. Hence, three human annotations (senior/junior resident orthopedic surgeon, and radiologist) can be found for this subset of data and the consistency between our model and human experts can be examined further.

2.2. Gaussian and Exponential Heatmaps

A Gaussian heatmap centering at ( x k , y k ) can be defined as:
G k ( x i , y j ;   σ )   =   ce ( x i x k ) 2   +   ( y i y k ) 2 2 σ 2 ,
where ( x i , y j ) is a pixel location of an image (on i-th column and j-th row), ( x k , y k ) is the human-annotated landmark location, σ is a hyperparameter that controls the width of the Gaussian, and c is the scaling constant, which was fixed to be a large constant (at the order of 10 3 ) so that the heatmap pixels close to landmark locations were more salient and easier to learn.
Representing landmark locations using heatmaps is beneficial. First, the regression of a heatmap was considered as an easier task as compared to the regression of x and y values, as the latter requires the learning of a highly non-linear function that maps all image pixels into two values only. Secondly, after normalization, the predicted heatmap is more informative, as it can be interpreted as a predicted probability map for the location of a landmark.
However, to determine G k ( x i , y j ;   σ ) , we need to select σ manually. The value of σ is often set as a small value because wide Gaussians may introduce high uncertainties to the prediction of landmark locations. However, if the Gaussian is not wide enough, features which are useful for prediction but outside the narrow region of Gaussian may have the chance to be ignored, as we simply let the model to predict zeros at the locations of those features.
To overcome this issue, we propose the use of exponential heatmaps for landmark localization. An exponential heatmap E k ( x i , y j ;   α ) centering at ( x k , y k ) can be defined as:
E k ( x i , y j ;   α )   =   ce α ( | x i x k | + | y j y k | ) ,
where ( x k , y k ) is the human-annotated landmark location, α is a hyperparameter that controls the decay rate of the exponential function, and c is the scaling constant as defined in the Gaussian heatmap. For the purpose of comparison, we let α =   1 σ log 2 / 2 so that both G k ( x i , y j ;   σ ) and E k ( x i , y j ;   σ ) reach to the same half maxima at ( x k   ± σ 2 log 2 , y k ) and ( x k ,   y k ± σ 2 log 2 ) , as shown in Figure 2.
Compared to the Gaussian heatmap G k ( x i , y j ;   σ ) , the proposed exponential heatmap E k ( x i , y j ;   σ ) has a sharper peak centering at the landmark location. As a result, the predicted locations of landmarks can be less uncertain. Furthermore, E k ( x i , y j ;   σ ) decays more slowly (especially when the pixels are close to y   =   y k and x   =   x k ), which leads to a wider region of heatmap pixels that should be predicted nonzero. In this case, useful features that lie within the non-zero zone are easier to detect and can be used as a clue for easier localization of landmarks.

2.3. Model Architecture

Our automatic SVA estimator (Figure 3) was based on a 31-layer ResUNet (Figure 4), which was composed of the head convolutional layers (C1), the downsampling path (C2–C5), upsampling path (C6–C9), and the tail convolutional layers (C10). In the downsampling path, both height and width of the feature maps were reduced by half four times. This process was then reversed during the upsampling path, and two heatmaps of landmarks were predicted immediately after C10.
The details of the ResUNet architecture are further illustrated in Figure 5 and Table 1. It is worth mentioning that, if the C7 center and the posterior corner of the sacrum were to be predicted on the same heatmap, then we would have to cope with landmark extraction and identification. For the former, we would have to, e.g., fit a Gaussian mixture model (which is rather complex) and regard the locations of the fitted Gaussian centers as the landmark locations. For the latter, we perhaps would have to assume that the one above is the landmark of C7, and the one below is the landmark of sacrum, as the extracted landmarks are indistinguishable.
In this work, all landmarks were predicted on separate heatmaps, i.e., the landmarks were separated and distinguishable from the very beginning. Thus, all we need to do was fit an exponential or a Gaussian function to each of the heatmaps and then regarded the obtained center of the fitted function as the estimated location of its corresponding landmark.

2.4. Image Preprocessing and Augmentation

Before training, all radiographs were downsized to 768 × 340 and then padded to 768 × 448 . The padding procedure was crucial, as the radiographs were randomly rotated (angle   [ 30 ° ,   30 ° ] ) during training, and the vertebrae part of the body should not be missing after rotation.
During training, techniques of data augmentation were applied to the radiographs in order to prevent CNNs from adapting to images of certain scales, orientations, and types of noise. The applied techniques were random scaling (scale [ 0.9 , 1.1 ] ), random rotation (angle   [ 30 ° ,   30 ° ] ), and random Gaussian blur (strength   [ 0 ,   0.5 ] ).

2.5. Details of Experiments

We trained a 31-layer ResUNet for the localization of the C7 center and the posterior corner of the sacrum. For model selection, the experimented hyperparameters for fine-tuning were heatmaps of different types (Gaussian and exponential) and sizes ( σ   =   3.14 ,   1.57 ,   0.78 ,   0.39 ). Furthermore, the procedure of 10-fold cross-validation was followed to test the effectiveness of the constructed models. For a schematic description about how the procedure of 10-fold cross-validation was performed, please refer to Figure S1.
Two GPUs were used for training (NVIDIA Tesla V100), and the batch size was set to 2 per GPU (therefore, the total batch size was 4). For the procedure of group normalization, the number of groups was set to 32. In addition, mean squared error was chosen as our loss function, which was minimized using stochastic gradient descent (SGD) with Nesterov momentum (momentum = 0.9). To eliminate the effect of overfitting, L2 regularization (strength = 10−5) was also added to the loss function.
During training, the learning rate was initially set to 0.01 and was then reduced by half on epoch 100. The number of training epochs was set to 120 and the final model was obtained from the checkpoint that performed the best on the validation data after epoch 80. The elapsed time of an experiment of 10-fold cross-validation was approximately 50 h. We used TensorFlow v1.12 and its Estimator API for model construction and training.

3. Results

We used absolute error of SVA and detection rate of landmarks to evaluate model performance. To obtain errors in pixels, we converted the predicted locations of landmarks back to the resolution of the original (unresized and unpadded) image and then calculated the pixel errors using locations of the true and predicted landmarks. The pixel errors were then converted to mm using the scaling factor recorded in each radiograph. We chose to represent errors in mm because scaling factor is different between radiographs and number of pixels would not be a universal standard.

3.1. Median Absolute Error of SVA

The median absolute error (MAE) of SVA is defined as:
MAE SVA   =   Median ( | SVA true ( 1 )     SVA pred . ( 1 ) | ,   | SVA true ( 2 )     SVA pred . ( 2 ) | ,   ... ,   | SVA true ( N )   SVA pred . ( N ) | ) ,
where N stands for number of samples. SVA true ( i ) and SVA pred . ( i ) stand for the i-th true and predicted SVA, respectively.
As MAE is insensitive to outliers, we also report the number of outliers (>10 mm) for absolute errors. The results of various hyperparameter settings are shown in Table 2.
Our results indicated that, when narrow Gaussian heatmaps were used ( σ   =   1.57 ,   0.79 ), the performance of SVA estimator became highly unstable (standard deviation > 10 mm). These results suggest that σ has to be fine-tuned carefully when choosing Gaussian heatmaps as the regressing targets.
To further investigate the quality of landmark localization, we also calculated the detection rate of the C7 center and the posterior corner of the sacrum, respectively. The results are shown in the following section.

3.2. Detection Rate of Landmarks

A landmark was considered to be detected when the distance between the true and predicted landmark locations was under some acceptable error threshold. To investigate the quality of predicted landmark locations, the detection rate was plotted against different degrees of error thresholds, as shown in Figure 6 (for the C7 center) and Figure 7 (for the posterior corner of the sacrum).
The results of detection rates indicate that the C7 center was difficult to be localized when narrow Gaussian heatmaps were used ( σ   =   1.57 ,   0.79 ). A possible reason is that there were few or no distinct features near the center of the vertebra. The distinct features, such as the vertebrae borders, were not included in the highlighted, non-zero region of the narrow Gaussian and were therefore difficult to be utilized for better localization of landmarks. By contrast, at small σ , the exponential function has not only a narrow region at its center, but also a rather extended region at its tail. Hence, for the exponential function, even if the distinct features were not inside the narrow region, they may still be discovered and utilized, if they were contained within the non-zero, extended tail region.
In contrast to models trained with narrow Gaussian heatmaps, models trained with narrow exponential heatmaps achieved the best performance for the detection of the C7 center. At σ   =   1.57 , the 1-mm and 5-mm detection rate of the C7 center reached 0.496 ± 0.023 and 0.911 ± 0.012, respectively, as shown in Table 3.
Additionally, we used fixed narrow heatmap function ( σ   =   1.57 ) and compared the 5-mm detection rate of models trained with different model complexity (31-layer ResUNet and 12-layer UNet) and different types of heatmaps (Gaussian and exponential). The results are shown in Table 4. Our results suggest that, when the model complexity was fixed, replacing the Gaussian heatmap with the exponential heatmap leads to a better and a more stable performance (the standard deviation has decreased by an order of magnitude). The 5-mm detection rate of C7 increased from 0.045 to 0.842, when a 12-layer UNet was used, and increased from 0.401 to 0.911, when a 31-layer ResUNet was used. These results indicate that the exponential heatmaps were easier to learn at σ   =   1.57 , regardless of the model complexity. Furthermore, when the heatmap type was fixed, replacing a 12-layer UNet with a 31-layer ResUNet led to significant improvements for the localization of the C7 center. This indicated that the extra skip-connections and convolutional layers of the 31-layer ResUNet were effective at σ   =   1.57 .

3.3. Analysis of Inter-Rater Reliability

To further compare the SVA estimates between ResUNet and human experts, or the SVA estimates between human experts only, we calculated the intra-class correlation coefficient (ICC). With ICC, the agreement between any two estimators can be tested. Our ICC scores were calculated by two-way-random effects model with absolute agreement. For the evaluation of ICC, 99 radiographs from one fold of 10-fold cross-validation were used.
We report the ICC scores and their 95% confidence intervals in Table 5. The ICC scores were evaluated for four participants, including a 31-layer ResUNet ( σ   =   1.57 , heatmap type = exponential), a junior resident orthopedic surgeon, a senior resident orthopedic surgeon, and a radiologist. Our results show that the ICC scores derived from any pair of the four participants were all >0.9, indicating a high agreement among the four participants.
We also report the Bland–Altman (B-A) plot for the SVA values estimated by the senior resident orthopedic surgeon and the ResUNet model in Figure 8. The median absolute difference and the mean difference of these two set of estimates was 1.34 mm and 0.32 mm, respectively. For a comparison between doctors, the B-A plot for the SVA values estimated by the senior and the junior resident orthopedic surgeons is shown in Figure 9. The median absolute difference, and the mean difference of these two set of estimates was 1.14 mm and 0.4 mm, respectively.

4. Discussion

Sagittal balance is a primary issue for clinical assessment of spine. There are several important morphological indicators of the alignment of spine and pelvis, including sagittal vertical axis (SVA), lumbar lordosis, pelvic incidence, pelvic tilt, and sacral slope. Among them, SVA is the most common parameter used for the evaluation of overall spinal alignment, which represents the sagittal curvature of the spine [25]. SVA can be used to determine the relationship between the spine and pelvis, and has a normal range. Yukawa et al. [26] reported the average value of 3.1 ± 12.6 mm for the SVA of 626 asymptomatic volunteers, and SVA may steadily increase from ages of 30 to 70 years.
Although SVA may change with aging, it remains a reliable sagittal spinopelvic parameter for the assessment of spine alignment. Several authors have demonstrated high inter-rater reliability of SVA measurement by the intra-class correlation coefficient (ICC). Kyrölä et al. [27] reported the ICC score of 0.99 (95% CI: 0.98, 1.00) for SVA in unselected adults with degenerative spine disorders. In a previous study, Katzman et al. [28] also showed a high ICC score of 0.93 (95% CI: 0.83, 0.97) in older adults with hyperkyphosis.
We selected the best model ( σ   =   1.57 , heatmap type = exponential) based on the absolute error of SVA. This model can be applied to patients of different age groups and with various clinical conditions, including patients with or without implants (Figure 10), having scoliosis, or degenerative changes (Figure 11).
The hyperparameter, σ , which controls the width of the heatmap function, plays an important role for model selection. In our experiments, we showed that the exponential heatmap allows the model to keep its performance and stability, when σ is small. This is potentially helpful when many landmarks exist. In that situation, fine-tuning σ values for each heatmap is very time-consuming and a heatmap function whose resulting performance is less sensitive to σ may be beneficial.
In this study, the inter-rater reliability of ResUNet and human experts was also examined. We classified the ICC scores similarly to the criteria introduced by Aubin and Kuklo et al. [29,30], i.e. ICC values greater than 0.90 indicate excellent reliability, values between 0.70 and 0.89 indicate good reliability, values between 0.50 and 0.69 indicate moderate reliability, values between 0.25 and 0.49 indicate low reliability, and values less than 0.24 indicate poor reliability. To the best of our knowledge, this study is the first to compare the inter-rater reliability between ResUNet and human experts, and obtained an excellent reliability with the range of ICC values from 0.946 to 0.993.
Using the ResUNet-based framework, the inference time was approximately 0.2 s per image (using one NVIDIA Tesla V100). Due to its short inference time and minimal labor requirement, the automated SVA estimator can be used as an SVA screener for either preoperative or postoperative patients. Nevertheless, recently, abundant important spinal parameters have been proposed and share the same importance as SVA, either at the coronal aspect or lateral aspect. Based on the success of this ResUNet-based SVA estimator, we would like to extend this robust deep learning method to other parameters in the future.
This study has some limitations. Firstly, the outline of vertebrae is slightly torched, as the X-ray was emitted from a single source of radiation. If, for example, the EOS™ X-ray machine [19] can be utilized, then the resulting outline of vertebrae, as well as the landmark locations predicted by our experts, may be more accurate. Secondly, when the selected CV model ( σ = 1.57, heatmap type = exponential) was used for predicting the entire dataset (990 images), we found 68 (6.9%) images were with large SVA absolute error (>10 mm). Among these 68 images, there were 33 and 17 images whose predictions were suspected to be influenced by numerical variation of mobile vertebrae (NVMV), and the occlusion of C7 or sacrum, respectively. In the former case, we found it difficult (for both human experts and for ResUNet) to predict the sacral point when confronting with cases of NVMV, or variable sacral fusion, by using the lateral plain radiographs only. NVMV is not uncommon in Chinese adults. In the study of Yan et al. [31], it was shown that 9.6% of patients have the condition of NVMV. We expect our model to handle patients with NVMV better, once more NVMV images are gathered and learned. In the latter case, by using the plain radiography, the C7 vertebral body was sometimes severely occluded by radiopaque areas around the shoulder, e.g., the clavicle, scapula, and humeral head, which may increase the difficulty for the algorithm to learn. By contrast, our human experts were generally able to infer the location of C7 body by observing the context of vertebrae and imagine the alignment of them.

5. Conclusions

The automatic measurement of spinal parameters is believed to have a great impact on orthopedics in the coming years. A fully automatic model was developed for the measurement of SVA using a ResUNet framework. The proposed model showed a median absolute error of 1.18 mm. The inference time for one image took approximately 0.2 s, which can be useful for fast screening of large datasets. Our model was cross-validated and fine-tuned to achieve optimal performance with a 31-layer ResUNet. The SVA values estimated by ResUNet was compared with the assessment of experienced doctors and showed excellent agreement. This proposed algorithm can also facilitate the process that allows orthopedic surgeons to design and plan the surgical procedures preoperatively and predict the postoperative outcomes. In addition, this method can be further improved by the ability to detect more landmark points for more clinical application in the future.

Supplementary Materials

The following are available online at https://www.mdpi.com/2077-0383/8/11/1826/s1, Figure S1: Schematic outline of the procedure of 10-fold cross-validation.

Author Contributions

Conceptualization, C.-Y.Y. and T.-T.T.; software, C.-H.W. and C.-Y.Y.; investigation, C.-H.W., C.-L.W., Y.-J.H. and Y.-C.Y.; data curation, C.-L.W., Y.-J.H., Y.-C.Y. and C.-J.F.; writing—original draft preparation, C.-H.W., C.-L.W., Y.-J.H. and Y.-C.Y.; writing—review and editing, T.-T.T.; and supervision, C.-Y.Y. and T.-T.T.

Funding

This study was supported by Chang Gung Memorial Hospital, Linkou (grant No. CORPG3J0401).

Acknowledgments

We are grateful to the National Center for High-Performance Computing for computer time and facilities. We’d like to acknowledgement the help of Center of Artificial Intelligence in Medicine, Chang Gung Memorial Hospital.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Djurasovic, M.; Glassman, S.D. Correlation of radiographic and clinical findings in spinal deformities. Neurosurg. Clin. N. Am. 2007, 18, 223–227. [Google Scholar] [CrossRef] [PubMed]
  2. D’Andrea, L.P.; Betz, R.R.; Lenke, L.G.; Clements, D.H.; Lowe, T.G.; Merola, A.; Haher, T.; Harms, J.; Huss, G.K.; Blanke, K.; et al. Do radiographic parameters correlate with clinical outcomes in adolescent idiopathic scoliosis? Spine 2000, 25, 1795–1802. [Google Scholar] [CrossRef] [PubMed]
  3. Glassman, S.D.; Berven, S.; Bridwell, K.; Horton, W.; Dimar, J.R. Correlation of radiographic parameters and clinical symptoms in adult scoliosis. Spine 2005, 30, 682–688. [Google Scholar] [CrossRef] [PubMed]
  4. Glassman, S.D.; Bridwell, K.; Dimar, J.R.; Horton, W.; Berven, S.; Schwab, F. The impact of positive sagittal balance in adult spinal deformity. Spine 2005, 30, 2024–2029. [Google Scholar] [CrossRef]
  5. Schwab, F.; Patel, A.; Ungar, B.; Farcy, J.P.; Lafage, V. Adult spinal deformity-postoperative standing imbalance: How much can you tolerate? An overview of key parameters in assessing alignment and planning corrective surgery. Spine 2010, 35, 2224–2231. [Google Scholar] [CrossRef]
  6. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. In Advances in Neural Information Processing Systems; Neural Information Processing Systems Foundation Inc.: San Diego, CA, USA, 2012; pp. 1097–1105. [Google Scholar]
  7. Litjens, G.; Kooi, T.; Bejnordi, B.E.; Setio, A.A.A.; Ciompi, F.; Ghafoorian, M.; Van Der Laak, J.A.; Van Ginneken, B.; Sánchez, C.I. A survey on deep learning in medical image analysis. Med. Image Anal. 2017, 42, 60–88. [Google Scholar] [CrossRef] [Green Version]
  8. Shen, D.; Wu, G.; Suk, H.-I. Deep learning in medical image analysis. Ann. Rev. Biomed. Eng. 2017, 19, 221–248. [Google Scholar] [CrossRef]
  9. Zeiler, M.D.; Fergus, R. Visualizing and understanding convolutional networks. In Computer Vision–ECCV 2014; Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T., Eds.; Springer: Cham, Switzerland, 2014; pp. 818–833. [Google Scholar]
  10. Cao, Z.; Hidalgo Martinez, G.; Simon, T.; Wei, S.E.; Sheikh, Y.A. OpenPose: Realtime Multi-Person 2D Pose Estimation using Part Affinity Fields. IEEE Trans. Pattern Anal. Mach. Intell. 2019. Available online: http://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=8765346&isnumber=4359286 (accessed on 27 October 2019).
  11. Colleoni, E.; Moccia, S.; Du, X.; De Momi, E.; Stoyanov, D. Deep Learning Based Robotic Tool Detection and Articulation Estimation With Spatio-Temporal Layers. IEEE Robot. Autom. Lett. 2019, 4, 2714–2721. [Google Scholar] [CrossRef] [Green Version]
  12. Drozdzal, M.; Vorontsov, E.; Chartrand, G.; Kadoury, S.; Pal, C. The importance of skip connections in biomedical image segmentation. In Deep Learning and Data Labeling for Medical Applications; Springer: Cham, Switzerland, 2016; pp. 179–187. [Google Scholar]
  13. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015. MICCAI 2015; Navab, N., Hornegger, J., Wells, W., Frangi, A., Eds.; Springer: Cham, Switzerland, 2015; pp. 234–241. [Google Scholar]
  14. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  15. Zhang, Z.; Liu, Q.; Wang, Y. Road extraction by deep residual u-net. IEEE Geosci. Remote Sens. Lett. 2018, 15, 749–753. [Google Scholar] [CrossRef]
  16. Lessmann, N.; van Ginneken, B.; de Jong, P.A.; Išgum, I. Iterative fully convolutional neural networks for automatic vertebra segmentation and identification. Med. Image Anal. 2019, 53, 142–155. [Google Scholar] [CrossRef] [Green Version]
  17. Wimmer, M.; Major, D.; Novikov, A.A.; Bühler, K. Fully automatic cross-modality localization and labeling of vertebral bodies and intervertebral discs in 3D spinal images. Int. J. Comput. Assist. Radiol. Surg. 2018, 13, 1591–1603. [Google Scholar] [CrossRef] [PubMed]
  18. Galbusera, F.; Niemeyer, F.; Wilke, H.J.; Bassani, T.; Casaroli, G.; Anania, C.; Costa, F.; Brayda-Bruno, M.; Sconfienza, L.M. Fully automated radiological analysis of spinal disorders and deformities: A deep learning approach. Eur. Spine. J. 2019, 28, 951–960. [Google Scholar] [CrossRef] [PubMed]
  19. Illés, T.; Somoskeöy, S. The EOS™ imaging system and its uses in daily orthopaedic practice. Int. Orthop. 2012, 36, 1325–1331. [Google Scholar] [CrossRef] [PubMed]
  20. Al Arif, S.M.R.; Knapp, K.; Slabaugh, G. Fully automatic cervical vertebrae segmentation framework for X-ray images. Comput. Meth. Prog. Biomed. 2018, 157, 95–111. [Google Scholar] [CrossRef] [Green Version]
  21. Wu, Y.; He, K. Group Normalization; Springer International Publishing: Cham, Switzerland, 2018; pp. 3–19. [Google Scholar]
  22. Chen, Y.; Wang, Z.; Peng, Y.; Zhang, Z.; Yu, G.; Sun, J. Cascaded Pyramid Network for Multi-Person Pose Estimation. Available online: http://openaccess.thecvf.com/content_cvpr_2018/html/Chen_Cascaded_Pyramid_Network_CVPR_2018_paper.html (accessed on 28 October 2019).
  23. Sun, K.; Xiao, B.; Liu, D.; Wang, J. Deep High-Resolution Representation Learning for Human Pose Estimation. Available online: https://arxiv.org/abs/1902.09212 (accessed on 28 October 2019).
  24. Pfister, T.; Charles, J.; Zisserman, A. Flowing convnets for human pose estimation in videos. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1913–1921. [Google Scholar]
  25. Decker, S.; Muller, C.W.; Omar, M.; Krettek, C.; Schwab, F.; Trobisch, P.D. Die sagittale Wirbelsäulenbalance – klinische Bedeutung und radiologische Bemessung. Z. Orthop. Unfall. 2016, 154, 128–133. [Google Scholar] [CrossRef]
  26. Yukawa, Y.; Kato, F.; Suda, K.; Yamagata, M.; Ueta, T.; Yoshida, M. Normative data for parameters of sagittal spinal alignment in healthy subjects: An analysis of gender specific differences and changes with aging in 626 asymptomatic individuals. Eur. Spine. J. 2018, 27, 426–432. [Google Scholar] [CrossRef]
  27. Kyrölä, K.K.; Salme, J.; Tuija, J.; Tero, I.; Eero, K.; Arja, H. Intra- and Interrater Reliability of Sagittal Spinopelvic Parameters on Full-Spine Radiographs in Adults with Symptomatic Spinal Disorders. Neurospine 2018, 15, 175–181. [Google Scholar]
  28. Katzman, W.B.; Parimi, N.; Gladin, A.; Fan, B.; Wong, S.S.; Mergenthaler, J.; Lane, N.E. Reliability of sagittal vertical axis measurement and association with measures of age-related hyperkyphosis. J. Phys. Ther. Sci. 2018, 30, 1417–1423. [Google Scholar] [Green Version]
  29. Aubin, C.E.; Bellefleur, C.; Joncas, J.; de Lanauze, D.; Kadoury, S.; Blanke, K.; Parent, S.; Labelle, H. Reliability and accuracy analysis of a new semiautomatic radiographic measurement soft-ware in adult scoliosis. Spine 2011, 36, 780–790. [Google Scholar] [CrossRef]
  30. Kuklo, T.R.; Potter, B.K.; O’Brien, M.F.; Schroeder, T.; Lenke, L.G.; Polly, D.; The Spinal Deformity Study Group. Reliability analysis for digital adolescent idiopathic scoliosis measurements. J Spinal. Disord Tech. 2005, 18, 152–159. [Google Scholar] [CrossRef]
  31. Yan, Y.-Z.; Li, Q.-P.; Wu, C.-C.; Pan, X.-X.; Shao, Z.-X.; Chen, S.-Q.; Wang, K.; Chen, X.-B.; Wang, X.-Y. Rate of presence of 11 thoracic vertebrae and 6 lumbar vertebrae in asymptomatic Chinese adult volunteers. J. Orthop. Surg. Res. 2018, 13, 124. [Google Scholar] [CrossRef] [PubMed] [Green Version]
Figure 1. Annotations of a radiograph. The C7 center and the posterior corner of the sacrum were annotated (red dots).
Figure 1. Annotations of a radiograph. The C7 center and the posterior corner of the sacrum were annotated (red dots).
Jcm 08 01826 g001
Figure 2. (AD) Exponential heatmaps at various σ values; and (EH) Gaussian heatmaps at various σ values. We put ( x k , y k ) at the heatmap center and draw the half-maximum contour of the heatmap (black-dashed lines) for each heatmap plot. When σ is fixed, the half-maxima contours of Gaussian and exponential heatmaps coincide at u 1   =   ( x k + σ 2 log 2 , y k ) , u 2   =   ( x k   σ 2 log 2 , y k ) , u 3   =   ( x k ,   y k + σ 2 log 2 ) and u 4   =   ( x k , y k σ 2 log 2 ) , as highlighted in (A,E).
Figure 2. (AD) Exponential heatmaps at various σ values; and (EH) Gaussian heatmaps at various σ values. We put ( x k , y k ) at the heatmap center and draw the half-maximum contour of the heatmap (black-dashed lines) for each heatmap plot. When σ is fixed, the half-maxima contours of Gaussian and exponential heatmaps coincide at u 1   =   ( x k + σ 2 log 2 , y k ) , u 2   =   ( x k   σ 2 log 2 , y k ) , u 3   =   ( x k ,   y k + σ 2 log 2 ) and u 4   =   ( x k , y k σ 2 log 2 ) , as highlighted in (A,E).
Jcm 08 01826 g002
Figure 3. Automatic sagittal vertical axis (SVA) estimator. This estimator will turn an input image (tenor shape =   768 × 448 × 1 ) into predictions of two probability maps of landmarks (tensor shape =   768 × 448 × 2 ), which correspond to the C7 center and the posterior corner of the sacrum. The two landmarks are then localized from their corresponding maps and their horizontal difference (i.e., SVA) is calculated.
Figure 3. Automatic sagittal vertical axis (SVA) estimator. This estimator will turn an input image (tenor shape =   768 × 448 × 1 ) into predictions of two probability maps of landmarks (tensor shape =   768 × 448 × 2 ), which correspond to the C7 center and the posterior corner of the sacrum. The two landmarks are then localized from their corresponding maps and their horizontal difference (i.e., SVA) is calculated.
Jcm 08 01826 g003
Figure 4. The ResUNet architecture. This network has 31 convolutional layers and is made of the following blocks: CONV, Res, Res-Down, Res-Down-C and Upsampling. The upsampling block performs bilinear upsampling and is non-learnable. The remaining four learnable blocks are illustrated in Figure 5.
Figure 4. The ResUNet architecture. This network has 31 convolutional layers and is made of the following blocks: CONV, Res, Res-Down, Res-Down-C and Upsampling. The upsampling block performs bilinear upsampling and is non-learnable. The remaining four learnable blocks are illustrated in Figure 5.
Jcm 08 01826 g004
Figure 5. Building blocks of the ResUNet: (a) CONV; (b) Res; (c) Res-Down/Res-Down-C. For all subfigures shown above, GN/ReLU stands for Group Normalization/ Rectified Linear Unit. In addition, CONV and Res were both used for feature extraction, where the latter improved information flow further by adding an extra skip-connection. Res-Down/Res-Down-C was used not only for feature extraction but also for data downsampling, where downsampling was achieved in the dimension of width and height (for Res-Down) or channel (for Res-Down-C). We set strides to 1 for the 1 × 1 Conv and the first 3 × 3 Conv in Res-Down. For all other convolutional layers, strides were set to 1.
Figure 5. Building blocks of the ResUNet: (a) CONV; (b) Res; (c) Res-Down/Res-Down-C. For all subfigures shown above, GN/ReLU stands for Group Normalization/ Rectified Linear Unit. In addition, CONV and Res were both used for feature extraction, where the latter improved information flow further by adding an extra skip-connection. Res-Down/Res-Down-C was used not only for feature extraction but also for data downsampling, where downsampling was achieved in the dimension of width and height (for Res-Down) or channel (for Res-Down-C). We set strides to 1 for the 1 × 1 Conv and the first 3 × 3 Conv in Res-Down. For all other convolutional layers, strides were set to 1.
Jcm 08 01826 g005
Figure 6. Detection rate of the C7 center against different degrees of error thresholds. The results of various σ values: (a) σ   =   6.28 ; (b) σ   =   3.14 ; (c) σ   =   1.57 ; and (d) σ   =   0.79 . The detection rates were results from 10-fold cross-validation (mean with 1 standard error band). For each figure, EXP and GS stand for exponential and Gaussian heatmaps, respectively.
Figure 6. Detection rate of the C7 center against different degrees of error thresholds. The results of various σ values: (a) σ   =   6.28 ; (b) σ   =   3.14 ; (c) σ   =   1.57 ; and (d) σ   =   0.79 . The detection rates were results from 10-fold cross-validation (mean with 1 standard error band). For each figure, EXP and GS stand for exponential and Gaussian heatmaps, respectively.
Jcm 08 01826 g006
Figure 7. Detection rate of the posterior corner of the sacrum against different degrees of error thresholds. The results of various σ values: (a) σ   =   6.28 ; (b) σ   =   3.14 ; (c) σ   =   1.57 ; and (d) σ   =   0.79 . The detection rates were results from 10-fold cross-validation (mean with 1 standard error band). For each figure, EXP and GS stand for exponential and Gaussian heatmaps, respectively.
Figure 7. Detection rate of the posterior corner of the sacrum against different degrees of error thresholds. The results of various σ values: (a) σ   =   6.28 ; (b) σ   =   3.14 ; (c) σ   =   1.57 ; and (d) σ   =   0.79 . The detection rates were results from 10-fold cross-validation (mean with 1 standard error band). For each figure, EXP and GS stand for exponential and Gaussian heatmaps, respectively.
Jcm 08 01826 g007
Figure 8. The Bland–Altman plot for the SVA values estimated by the senior resident orthopedic surgeon and the ResUNet model. The horizontal lines of mean difference (solid line) and mean difference ± 1.96 standard deviation (SD) are plotted (dashed lines).
Figure 8. The Bland–Altman plot for the SVA values estimated by the senior resident orthopedic surgeon and the ResUNet model. The horizontal lines of mean difference (solid line) and mean difference ± 1.96 standard deviation (SD) are plotted (dashed lines).
Jcm 08 01826 g008
Figure 9. The Bland–Altman plot for the SVA values estimated by the senior and the junior resident orthopedic surgeons. The horizontal lines of mean difference (solid line) and mean difference ± 1.96 standard deviation (SD) are plotted (dashed lines).
Figure 9. The Bland–Altman plot for the SVA values estimated by the senior and the junior resident orthopedic surgeons. The horizontal lines of mean difference (solid line) and mean difference ± 1.96 standard deviation (SD) are plotted (dashed lines).
Jcm 08 01826 g009
Figure 10. Examples of predicted images of post spinal fusion surgery with varied length of implants: (a) thoracic-lumbar fusion; (b) thoracic fusion; and (c) lumbar-sacral fusion.
Figure 10. Examples of predicted images of post spinal fusion surgery with varied length of implants: (a) thoracic-lumbar fusion; (b) thoracic fusion; and (c) lumbar-sacral fusion.
Jcm 08 01826 g010
Figure 11. Examples of predicted images of different clinical conditions. These images include patients of: (a) ages 8; (b) age 23; and (c) age 63, who also present with degenerative change.
Figure 11. Examples of predicted images of different clinical conditions. These images include patients of: (a) ages 8; (b) age 23; and (c) age 63, who also present with degenerative change.
Jcm 08 01826 g011
Table 1. The architecture of a 31-layer ResUNet.
Table 1. The architecture of a 31-layer ResUNet.
LevelBlock# 3 × 3
Conv
# 1 × 1
Conv
Output Shape
( N   × H   × W   ×   C )
C1CONV10N × 768 × 448 × 64
Res20N × 768 × 448 × 64
C2Res-Down21N × 384 × 224 × 128
Res20N × 384 × 224 × 128
C3Res-Down21N × 192 × 112 × 256
C4Res-Down21N × 96 × 56 × 512
C5Res-Down21N × 48 × 28 × 512
Res20N × 48 × 28 × 512
C6Upsampling00N × 96 × 56 × 512
C7Res-Down-C21N × 96 × 56 × 256
Upsampling00N × 192 × 112 × 256
C8Res-Down-C21N × 192 × 112 × 128
Upsampling00N × 384 × 224 × 128
C9Res-Down-C21N × 384 × 224 × 64
Upsampling00N × 768 × 448 × 64
C10CONV10N × 768 × 448 × 64
CONV10N × 768 × 448 × 32
CONV10N × 768 × 448 × 2
N, Number of images passed into the architecture; H/W/C, Height/Width/Channel of the input images; # 3 × 3 Conv, Number of 3 × 3 convolutional layers; # 1 × 1 Conv, Number of 1 × 1 convolutional layers.
Table 2. The median value and number of outliers of the absolute errors of SVA are reported for models constructed using different hyperparameter settings. Mean ± standard deviation (SD) of the median value of the absolute errors of SVA were obtained following the procedure of 10-fold cross-validation.
Table 2. The median value and number of outliers of the absolute errors of SVA are reported for models constructed using different hyperparameter settings. Mean ± standard deviation (SD) of the median value of the absolute errors of SVA were obtained following the procedure of 10-fold cross-validation.
Heatmap TypeHeatmap Width (σ)Median Absolute Error of SVA (Mean ± SD, mm)Number of Outliers
(>10 mm)
EXP6.281.570 ± 0.26867 (6.77%)
GS6.281.508 ± 0.30554 (5.45%)
EXP3.141.222 ± 0.16269 (6.97%)
GS3.141.206 ± 0.13962 (6.26%)
EXP1.571.183 ± 0.16668 (6.87%)
GS1.579.826 ± 14.1293 (29.6%)
EXP0.791.210 ± 0.12382 (8.28%)
GS0.7928.80 ± 10.4588 (59.4%)
EXP, Exponential heatmap; GS, Gaussian heatmap.
Table 3. The 1-mm and 5-mm detection rates of models trained using different hyperparameter settings. Mean ± standard deviation (SD) of the detection rates were obtained following the procedure of 10-fold cross-validation.
Table 3. The 1-mm and 5-mm detection rates of models trained using different hyperparameter settings. Mean ± standard deviation (SD) of the detection rates were obtained following the procedure of 10-fold cross-validation.
Heatmap TypeHeatmap Width (σ)DC7 at 1 mm
Mean ± SD
DC7 at 5 mmDS at 1 mmDS at 5 mm
Mean ± SDMean ± SDMean ± SD
EXP6.280.279 ± 0.0130.863 ± 0.0150.220 ± 0.0140.859 ± 0.011
GS6.280.313 ± 0.0270.858 ± 0.0110.235 ± 0.0230.867 ± 0.0093
EXP3.140.433 ± 0.0170.899 ± 0.00460.289 ± 0.0290.872 ± 0.0054
GS3.140.460 ± 0.0160.898 ± 0.00930.309 ± 0.0280.853 ± 0.015
EXP1.570.496 ± 0.0230.911 ± 0.0120.328 ± 0.0160.871 ± 0.0037
GS1.570.231 ± 0.130.401 ± 0.230.354 ± 0.0180.849 ± 0.026
EXP0.790.487 ± 0.0260.873 ± 0.0120.349 ± 0.0120.827 ± 0.016
GS0.790.012 ± 0.0500.018 ± 0.0390.307 ± 0.0130.792 ± 0.026
EXP, Exponential heatmap; GS, Gaussian heatmap; DC7, Detection rate of the C7 center; DS, Detection rate of the posterior corner of the sacrum.
Table 4. The 5-mm detection rates of models trained using σ   =   1.57 . Mean ± standard deviation (SD) of the detection rates were obtained following the procedure of 10-fold cross-validation.
Table 4. The 5-mm detection rates of models trained using σ   =   1.57 . Mean ± standard deviation (SD) of the detection rates were obtained following the procedure of 10-fold cross-validation.
ModelNumber of LayersHeatmap TypeDC7 at 5 mmDS at 5 mm
Mean ± SDMean ± SD
UNet12GS0.045 ± 0.0780.750 ± 0.039
UNet12EXP0.842 ± 0.00790.853 ± 0.0093
ResUNet31GS0.401 ± 0.230.849 ± 0.026
ResUNet31EXP0.911 ± 0.0120.871 ± 0.0037
EXP, Exponential heatmap; GS, Gaussian heatmap; DC7, Detection rate of the C7 center; DS, Detection rate of the posterior corner of the sacrum.
Table 5. The intra-class correlation coefficient (ICC) and its 95% confidence interval (CI). The ICC values were evaluated for different pair of estimators.
Table 5. The intra-class correlation coefficient (ICC) and its 95% confidence interval (CI). The ICC values were evaluated for different pair of estimators.
Measurement: SVA (mm)ICC with 95% CI 95% CI
ResUNet vs. Rater 1 10.9890.984–0.993
ResUNet vs. Rater 2 20.9460.920–0.963
ResUNet vs. Rater 3 20.9930.989–0.995
Rater 1 vs. Rater 20.9490.925–0.966
Rater 1 vs. Rater 30.9960.993–0.997
Rater 2 vs. Rater 30.9550.934–0.970
1 Rater 1, Radiologist. 2 Rater 2 (Rater 3), Junior (Senior) resident orthopedic surgeon.

Share and Cite

MDPI and ACS Style

Weng, C.-H.; Wang, C.-L.; Huang, Y.-J.; Yeh, Y.-C.; Fu, C.-J.; Yeh, C.-Y.; Tsai, T.-T. Artificial Intelligence for Automatic Measurement of Sagittal Vertical Axis Using ResUNet Framework. J. Clin. Med. 2019, 8, 1826. https://doi.org/10.3390/jcm8111826

AMA Style

Weng C-H, Wang C-L, Huang Y-J, Yeh Y-C, Fu C-J, Yeh C-Y, Tsai T-T. Artificial Intelligence for Automatic Measurement of Sagittal Vertical Axis Using ResUNet Framework. Journal of Clinical Medicine. 2019; 8(11):1826. https://doi.org/10.3390/jcm8111826

Chicago/Turabian Style

Weng, Chi-Hung, Chih-Li Wang, Yu-Jui Huang, Yu-Cheng Yeh, Chen-Ju Fu, Chao-Yuan Yeh, and Tsung-Ting Tsai. 2019. "Artificial Intelligence for Automatic Measurement of Sagittal Vertical Axis Using ResUNet Framework" Journal of Clinical Medicine 8, no. 11: 1826. https://doi.org/10.3390/jcm8111826

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop