Next Article in Journal
Analytical Performance and Concordance with Next-Generation Sequencing of a Rapid, Multiplexed dPCR Panel for the Detection of DNA and RNA Biomarkers in Non-Small-Cell Lung Cancer
Previous Article in Journal
Evaluation of an Isothermal Amplification HPV Assay on Self-Collected Vaginal Samples as Compared to Clinician-Collected Cervical Samples
Previous Article in Special Issue
Reversible Cerebral Vasoconstriction Syndrome and Raynaud’s Phenomenon: Is There a Link between the Pathogeneses of Their Underlying Complex Etiology? A Case Report and Literature Review
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deep Learned Segmentations of Inflammation for Novel ⁹⁹mTc-maraciclatide Imaging of Rheumatoid Arthritis

1
Department of Biomedical Engineering, School of Biomedical Engineering and Imaging Sciences, King’s College London, London WC2R 2LS, UK
2
Department of Cancer Imaging, School of Biomedical Engineering and Imaging Sciences, King’s College London, London WC2R 2LS, UK
3
King’s College London and Guy’s and St Thomas’ PET Centre, King’s College London, London WC2R 2LS, UK
*
Author to whom correspondence should be addressed.
Diagnostics 2023, 13(21), 3298; https://doi.org/10.3390/diagnostics13213298
Submission received: 8 September 2023 / Revised: 4 October 2023 / Accepted: 10 October 2023 / Published: 24 October 2023
(This article belongs to the Special Issue Imaging and Artificial Intelligence in Rheumatology)

Abstract

:
Rheumatoid arthritis (RA) is an autoimmune disease that causes joint pain, stiffness, and erosion. Power Doppler ultrasound and MRI are imaging modalities used in detecting and monitoring the disease, but they have limitations. ⁹⁹mTc-maraciclatide gamma camera imaging is a novel technique that can detect joint inflammation at all sites in a single examination and has been shown to correlate with power Doppler ultrasound. In this work, we investigate if machine learning models can be used to automatically segment regions of normal, low, and highly inflamed tissue from 192 ⁹⁹mTc-maraciclatide scans of the hands and wrists from 48 patients. Two models were trained: a thresholding model that learns lower and upper threshold values and a neural-network-based nnU-Net model that uses a convolutional neural network (CNN). The nnU-Net model showed 0.94 ± 0.01, 0.51 ± 0.14, and 0.76 ± 0.16 modified Dice scores for segmenting the normal, low, and highly inflamed tissue, respectively, when compared to clinical segmented labels. This outperforms the thresholding model, which achieved modified Dice scores of 0.92 ± 0.01, 0.14 ± 0.07, and 0.35 ± 0.21, respectively. This is an important first step in developing artificial intelligence (AI) tools to assist clinicians’ workflow in the use of this new radiopharmaceutical.

1. Introduction

Currently, ultrasound and MRI are widely used to detect and monitor inflamed tissue in rheumatoid arthritis (RA). 99mTc-maraciclatide imaging has been shown to correlate well with power Doppler ultrasound [1] and is being investigated as an adjunct to these imaging modalities. Unlike MRI, 99mTc-maraciclatide is associated with ionizing radiation but is widely available and faster than MRI and can scan all joints in a single scan, unlike ultrasound, where each joint is scanned individually. These other modalities have been used to monitor RA for some time, and there are studies on the use of artificial intelligence (AI) models to help with early detection [2] and monitoring of the disease [3]. This includes early grading of RA using MRI [4] and classifying metacarpophalangeal joints using ultrasound images [5].
AI tools are increasingly being incorporated into clinicians’ workflows to help with clinical efficiency [6] and improve interobserver agreement [7]. Image segmentation is one such task that has been studied extensively [8,9,10], with various methods and tools, such as data augmentation with generative adversarial networks [11,12], cascade networks [13], and deep supervision [14], that can be used to improve performance.
In terms of AI imaging diagnostics in RA, there have been several recent papers on the topic, such as [15] where Alarcón-Paredes et al. used thermal and RGB images as well as other collected features (weight, height, age, etc.) from female patients to determine if the patient has RA. There are also reports [16,17] in which RA diagnosis is determined from a convolutional neural network (CNN) based on hand X-ray images. In [18], similar to [15], RGB images were used with questionnaire information to diagnose RA.
In this study, our aim is to build AI tools for 99mTc-maraciclatide imaging to detect inflamed tissue in RA. Our hypothesis is that machine learning models can be used to automatically segment normal and inflamed tissue of patients’ hands (Figure 1 and Figure 2) with RA.

2. Materials and Methods

A thresholding and nnU-Net model were trained and compared. The overview of the methodology for a single crossfold can be seen in Figure 3.

2.1. Data

The data were collected as part of a previous study and consisted of 192 hand and wrist images (images were of the size 256 × 256) from 48 patients, including palmar, dorsal, and two oblique views (Figure 1). Each patient had an injection of 752 ± 32 MBq of 99mTc-maraciclatide, and after a two-hour period, the hand views (as well as feet and whole-body views) were acquired using a gamma camera over a 1 h scanning period. The images were then manually segmented into non-mutually exclusive regions of normal, low, and high inflammation by a clinician (GC) experienced with 99mTc-maraciclatide imaging in RA. To make the labels mutually exclusive, i.e., one classification per pixel, the more severe class was taken. For example, a pixel that was classified as both high and low inflammation was marked as only highly inflamed, and pixels that were classified as normal and low inflammation were marked as low inflammation. Any pixels that were not classified as normal, low, or high inflammation were marked as background. The images were segmented using label-studio [19], where the images were exported with fixed color scaling from 1 to 100 based on the clinician’s decision after being shown a selection of images for different color scales (Figure 2).

2.2. Thresholding

To serve as a baseline model, we trained a simple thresholding model to segment regions of normal, low, and highly inflamed tissue. The thresholding model consisted of first smoothing the input image using a Gaussian kernel followed by lower and upper threshold values of the smoothed image (Figure 4). The width of the Gaussian kernel is characterized by the sigma value ( σ ), and the neighborhood size of the kernel in terms of pixels is based on Equation (1).
k e r n e l _ s i z e = r o u n d ( max 3 σ ,   1 ) + 1 i f   r o u n d ( max 3 σ ,   1 )   i s   e v e n   r o u n d max 3 σ ,   1 o t h e r w i s e
Each of the three labels, normal, low, and high inflammation, was learned independently, leading to a model that consists of only nine trainable parameters (a sigma value, lower and upper threshold values for each of the three labels).
The sigma, lower, and upper threshold values were selected by grid search over the training dataset. The search values of σ were from 0.1 to 4.0 every 0.1, and for the lower and upper threshold values, the search was in the range from 0 to the maximum value in the training dataset. The dice score was used to select the model parameters that performed the best over the validation dataset. The thresholding parameters that performed the best over the validation dataset in the inner crossfold were then tested on the test dataset; more information on the crossfolding can be found in Section 2.8.

2.3. nnU-Net

We also used the nnU-Net [20] model to segment the regions of inflammation in the images. The nnU-Net model is a convolutional neural network biomedical imaging segmentation framework. It uses a U-Net [21] style convolutional neural network with instance normalization and leakyReLU activation functions. It uses a stochastic gradient descent (SGD) optimizer with a combined Dice and cross-entropy loss trains for 1000 epochs with a polynomial learning rate scheduler (a scheduler that decreases the learning rate in accordance with a polynomial function). The nnU-Net model uses data augmentation strategies such as rotations and adding Gaussian noise. The nnU-Net framework trains three models: a 2D U-Net, a 3D U-Net, and a (3D) cascade U-Net; however, due to our data only being 2D, only the 2D U-Net model is trained and evaluated. The nnU-Net model version used was the version 1 model, and the code can be found in the GitHub repository: https://github.com/MIC-DKFZ/nnUNet/tree/nnunetv1 (accessed on 13 March 2023); the trained weights are available on request.

2.4. Intraobserver Variability

In addition to the two models being compared, the clinician who segmented the data also re-segmented 15 % of the data a month after the initial data were segmented using the same labeling procedure outlined in Section 2.1. The re-segmented images were compared to the original segmentations using the same metrics to give a comparison of how the AI models compare to a human observer. Standard deviation numbers are based on crossfolds and are therefore not available for the intraobserver data.

2.5. Dice Score

A modified Dice score was used for the evaluation of the models. The regular Dice score is defined in Equation (2). Due to the dataset having a large number of empty segmentation maps (56 images have no low inflammation, 87 images have no high inflammation, and 45 images have neither low nor high inflammation) and the known issues that the Dice score has with empty segmentations as discussed in [22], we used a modified version of the Dice score (Equation (3)) to define the comparison of two empty segmentation maps to give a modified Dice score of 1 as opposed to undefined in the initial formulation. The Dice score from the MONAI [23] data package was used for the modified Dice metric.
D i c e   S c o r e = 2 X Y X + Y
M o d i f i e d   D i c e   S c o r e = 2 X Y X + Y   i f   X + Y   0 1 o t h e r w i s e

2.6. Intersection over Union (IoU)

A modified IoU score was also used for the evaluation of the models. Similar to the modified dice score above, the regular IoU (Equation (4)) score is modified to account for empty segmentation comparison (Equation (5)). The IoU metric from the MONAI [23] data package was used for the modified IoU metric.
I o U   S c o r e = X Y   X Y
M o d i f i e d   I o U   S c o r e = X Y   X Y   i f   X   Y   0 1 o t h e r w i s e

2.7. Confusion Matrix

To further analyze the results of the model, confusion matrices were used to show the breakdown of the proportion of pixels classified correctly and incorrectly, including what misclassified pixels were classified as. These matrices were obtained for both trained models as well as for analyzing the intraobserver variability. The model confusion matrices were calculated over all outer crossfolds, and the intraobserver confusion matrix was taken over the ~15% of data that was re-segmented.

2.8. Crossfolding

To limit bias in the assessment of our models, we used crossfolding to create 10 models for the nnU-Net model as well as for the thresholding model. The dataset was partitioned into 10 buckets, each bucket consisting of 4 or 5 patients so that no patient images were shared across a bucket. One bucket was reserved as a held-out test dataset. The other nine buckets were used as a train/validation dataset (Figure 5). Each image within the overall dataset exists in the test dataset for one crossfold and in the train/validation dataset for nine crossfolds. Results presented show means ± standard deviations based on these crossfolds. A box-whisker plot is shown using the variance in results presented using different crossfolds.
For the thresholding model, an inner crossfold was also used. For each outer crossfold, we trained nine models, one for each of the nine train/validation buckets being held as a validation dataset. The thresholding model was trained on eight buckets of data, and the model that performed best on the validation bucket was then tested using the held-out test dataset so that 10 models in total were tested on the test dataset.

2.9. Single Inflammation Class

The model segmentations and the labels split the inflammation into two classes: low and high inflammation. The two classes can be combined into one “inflamed” class by adding the labels together. The mutually exclusive constraint introduced prior ensures that the addition of the labels results in a new valid label, as no pixel is marked as both low and high inflammation. The models can then be compared with respect to this combined class. The models still make two separate segmentations for both low and high inflammation, but the two classes are then combined and analyzed. Note that this is different from training two additional models that are trained to output a single inflamed class. (Figure 6).

2.10. ROC Curves

We also retroactively analyzed our generated segmentation maps as a classification task by classifying an image as inflamed if the corresponding label contains at least one pixel of inflammation (high or low). Using this, we created a ROC curve using the number of pixels classified as inflammation in the prediction as our variable threshold. For example, when 0 inflamed pixels are needed to classify an image as inflamed, we gain a true positive rate and false positive rate of 1; when the number of pixels needed to classify the image as inflamed is equal to the entire number of pixels in the image, we get a true positive rate and false positive rate of 0.

3. Results

Box-Wisker plots of the nnU-Net and thresholding models are shown in Figure 7. The nnU-Net model segmented inflammation with modified Dice scores of 0.94 ± 0.01, 0.51 ± 0.14, and 0.76 ± 0.16 for normal, low inflamed and highly inflamed tissue. The thresholding model, in contrast, gave modified Dice scores of 0.92 ± 0.01, 0.14 ± 0.07, and 0.35 ± 0.21 (Table 1). The results of the nnU-Net model are comparable to the results of the intraobserver-modified Dice scores of 0.94, 0.51, and 0.63.
We also computed modified IoU scores for the nnU-Net as 0.89 ± 0.01, 0.43 ± 0.15, and 0.70 ± 0.15 for the normal, low, and high channels, respectively. For the thresholding model, the modified IoU results were 0.85 ± 0.02, 0.11 ± 0.07, and 0.28 ± 0.18 for normal, low, and high inflammation, respectively.
The confusion matrices (Figure 8) show that the nnU-Net model accurately classifies over 99% of background pixels, over 94% of normal tissue pixels, and approximately 39% of low and 82% of highly inflamed tissue pixels. For the low inflammation class, where the model performance is weakest, the model predicts the low inflammation tissue as normal tissue ~52% of the time and predicts it as highly inflamed ~9% of the time. Interestingly, when the model inaccurately classifies highly inflamed tissue, it does so as normal tissue (~15%) more than it does low inflamed (~3%). The thresholding confusion matrix, in comparison, shows less accuracy for all classes. The thresholding model misclassifies high inflammation, mostly as low inflammation, as one would expect, as opposed to the nnU-Net model, which mostly misclassifies high inflammation as normal tissue. Lastly, a confusion matrix for the intraobserver variability is presented. This shows that the intraobserver variance is relatively large for regions of low and high inflammation.
Looking at some example segmentations, we can see the thresholding model predicting several areas of small amounts of inflammation all over the hand image as the decision made using the thresholding is relatively local. The nnU-Net model, in contrast, shows a smaller number of large areas of inflammation being predicted, more in line with the labels. The examples also show that the nnU-Net model is penalized in the modified Dice score for errors in boundary prediction (Figure 9).

3.1. Inflammation as a Single Class

In addition to looking at high and low inflammation separately, we also evaluated the model performance when combining these two classes into one “inflammation” class. If the two classes predicted by the nnU-Net model are combined into a single inflamed class and compared to the combined inflammation in the clinically segmented labels, we get a modified Dice score of 0.72 ± 0.12. This, again, is broadly in line with the intraobserver modified Dice score of 0.76 when collapsing the inflammation in the same manner. The score of the inflamed channel for the nnU-Net model (0.72 ± 0.12) is much closer to the model’s high-inflammation-modified Dice score of 0.76 ± 0.16 than the low-modified Dice score of 0.51 ± 0.14. This is in part due to the fact that mistakes in inflammation classification are not counted against the model in this way of analyzing the data (i.e., errors of misclassifying low inflammation as high inflammation and vice versa are ignored when just considering inflammation as a single class). This could also be due to the relative sizes of high inflammation and low inflammation in the data; approximately 0.42% of all the pixels are classified as highly inflamed compared to 0.25% as low inflammation.
The thresholding model, in comparison, gave results of 0.40 ± 0.18 when analyzed in this manner. This significantly underperformed in comparison to the nnU-Net model and shows again that the combined inflammation is much closer to the higher modified Dice score of 0.35 ± 0.21 than 0.14 ± 0.07 given by the high inflammation than the low inflammation. Table 2 shows the results of combining the class for each crossfold for the two models.

3.2. ROC Curve

The ROC curve is shown in Figure 10, and the calculated area under the curve (AUC) is 0.96 for the nnU-Net model and 0.88 for the thresholding model. For the case of using 1 pixel in the prediction as our threshold, we gain a sensitivity of 0.98 ± 0.06 and specificity of 0.80 ± 0.31 for the nnU-Net model and a sensitivity of 1.00 ± 0.00 and specificity of 0.22 ± 0.36 for the thresholding model.

4. Discussion

This study is the first to investigate automatic segmentation of inflammation in RA patients with 99mTc-maraciclatide; as such, direct comparisons are hard to find in the literature. There are studies in the literature that have automatically tried to classify arthritic disease activity via Color Doppler ultrasound [24] and segment regions of inflamed tissue with MRI [25]. In [24], the authors report an accuracy rate of 75.0% when predicting an OMERACT-EULAR Synovitis Scoring (OESS) score over the entire US image and 87% when predicting healthy vs. diseased. For the healthy/disease model, they also report sensitivity and specificity of 0.864 and 0.875. Comparing these numbers to our calculated sensitivity and specificity for the nnU-Net model shows we have a higher sensitivity rate (0.98 vs. 0.864) but lower specificity (0.80 vs. 0.875). They also use a larger dataset of 1342 images, which can greatly affect the performance of machine learning models. The research presented here differs in that we are segmenting instead of classifying and using a completely different modality.

4.1. Relative Performance of Low and High Inflammation

Both models and the intraobserver comparison showed better performance at segmenting high inflammation than low inflammation. When looking at the confusion matrices, it seems that low inflammation tissue is confused for normal tissue more than it is high inflammation. This could partly be because of the lack of examples of low inflammation in the training dataset relative to high inflammation, as mentioned previously (0.42% of pixels for high and 0.25% of pixels for low). However, in addition to this pixel level imbalance, if we analyze the average amount of pixels in an ‘inflamed’ lesion, we get 62.81 ± 82.55 pixels for low and 121.25 ± 130.65 pixels for high inflammation lesions. This shows that not only are there fewer pixels of low inflammation to learn from but that, on average, the high inflammation is split into a larger number of smaller lesions. This could also contribute to the low intraobserver modified Dice score, as boundary imperfections will have a larger impact than drawing larger, fewer lesions.

4.2. Imbalanced Dataset

There have been several attempts in the literature to create loss functions that help with training on imbalanced datasets, such as focal loss [26] that modifies cross-entropy loss to help handle class imbalance and Tversky loss that has been shown to outperform Dice loss in some tasks [27].
The nnU-Net model uses traditional augmentation, such as Gaussian blurring and rotations, to boost the train dataset size. There has been widespread adoption of generative augmentation techniques such as generative adversarial networks (GANs) [28] and, more recently, diffusion models [29] that could be used to augment the dataset with synthetic samples. Papers using synthetic generation have, in some cases, shown significantly improved results [30,31,32] and could be used here to improve the results of our model, especially with the low, high inflammation imbalance similar to [33].

4.3. Limitations

The research presented here currently only predicts clinically segmented labels, which, in this instance, have demonstrably high variability at the pixel level. The work is also based on the Dice score, which, whilst a standard metric in machine learning and medical imaging, has known issues with empty segmentations; it is also questionable if such a metric is suited to our task as the exact boundaries of the inflamed tissue are less clinically relevant than simply predicting if a specific joint has any inflammation. In addition, our labels are based on a single observer, and a better understanding of the data may arise from looking at predictions from multiple observers. Future work is needed to investigate these issues.
The work presented here is also limited in that the dataset is relatively small and is collected from only one center and thus may not be representative of data collected from other geographical sites/regions. Further work is warranted to study the generalisability of the models presented here.

5. Conclusions

In conclusion, two machine learning models were trained and compared on the task of segmenting normal tissue and regions of inflammation in the hands of patients with RA using 99mTc-maraciclatide imaging. The nnU-Net model shows a promising ability to segment regions of low and highly inflamed tissue, with similar performance to a human observer in terms of repeatability for segmenting normal and highly inflamed tissue. The nnU-Net model also outperformed a thresholding model built for this task. Further work may include getting multiple trained observers to segment the images and study the interobserver variance, as well as using these multiple labels to learn a more robust model and quantify uncertainty. We believe this work is the first step in building a fast and reliable clinical assistance pipeline to use this new radiopharmaceutical in conjunction with other modalities to improve efficiency in detecting and monitoring synovitis in RA patients.

Author Contributions

Conceptualization, R.C., G.J.R.C. and A.J.R.; methodology, R.C., G.J.R.C. and A.J.R.; software, R.C.; validation, R.C.; formal analysis, R.C.; investigation, R.C.; resources, R.C., G.J.R.C. and A.J.R.; data curation, R.C., G.J.R.C.; writing R.C., G.J.R.C. and A.J.R.; writing—review and editing, R.C., G.J.R.C. and A.J.R.; visualization, R.C.; supervision, G.J.R.C. and A.J.R.; project administration, G.J.R.C. and A.J.R.; funding acquisition, G.J.R.C. and A.J.R. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded in whole, or in part, by the Wellcome Trust [WT203148/Z/16/Z]. For the purpose of open access, the author has applied a CC BY public copyright licence to any Author Accepted Manuscript version arising from this submission. The Wellcome Trust has not had any implications regarding the conceptualisation of this research. This research was funded by Serac Healthcare.

Institutional Review Board Statement

The data for the study was collected in accordance with the Declaration of Helsinki and approved by the NHS HRA on 7 April 2016 (IRAS ID: 188145 REC: 16/LO/0309). Institutional approval was obtained for the analysis of the anonymized data.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

Data is provided by Guy’s and St Thomas’s NHS Trust and, due to privacy and ethical reasons, cannot be shared at this time.

Conflicts of Interest

G.J.R.C. has received funding from Serac Healthcare for research projects and has acted as a consultant in the past for Serac Healthcare.

References

  1. Attipoe, L.; Chaabo, K.; Wajed, J.; Hassan, F.U.; Shivapatham, D.; Morrison, M.; Ballinger, J.; Cook, G.; Cope, A.P.; Garrood, T. Imaging neoangiogenesis in rheumatoid arthritis (INIRA): Whole-body synovial uptake of a 99mTc-labelled RGD peptide is highly correlated with power Doppler ultrasound. Ann. Rheum. Dis. 2020, 79, 1254–1255. [Google Scholar] [CrossRef] [PubMed]
  2. Stoel, B.C. Artificial intelligence in detecting early RA. In Seminars in Arthritis and Rheumatism; WB Saunders: Philadelphia, PA, USA, 2019; Volume 49, pp. S25–S28. [Google Scholar]
  3. Momtazmanesh, S.; Nowroozi, A.; Rezaei, N. Artificial intelligence in rheumatoid arthritis: Current status and future perspectives: A state-of-the-art review. Rheumatol. Ther. 2022, 9, 1249–1304. [Google Scholar] [CrossRef] [PubMed]
  4. More, S.; Singla, J. A Study on Automated Grading System for Early Prediction of Rheumatoid Arthritis. In Proceedings of the 2021 6th International Conference on Communication and Electronics Systems (ICCES), Coimbatre, India, 8–10 July 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 1293–1300. [Google Scholar]
  5. Wu, M.; Wu, H.; Wu, L.; Cui, C.; Shi, S.; Xu, J.; Liu, Y.; Dong, F. A deep learning classification of metacarpophalangeal joints synovial proliferation in rheumatoid arthritis by ultrasound images. J. Clin. Ultrasound 2022, 50, 296–301. [Google Scholar] [CrossRef] [PubMed]
  6. Chua, I.S.; Gaziel-Yablowitz, M.; Korach, Z.T.; Kehl, K.L.; Levitan, N.A.; Arriaga, Y.E.; Jackson, G.P.; Bates, D.W.; Hassett, M. Artificial intelligence in oncology: Path to implementation. Cancer Med. 2021, 10, 4138–4149. [Google Scholar] [CrossRef] [PubMed]
  7. Brinjikji, W.; Abbasi, M.; Arnold, C.; Benson, J.C.; Braksick, S.A.; Campeau, N.; Carr, C.M.; Cogswell, P.M.; Klaas, J.P.; Liebo, G.B.; et al. e-ASPECTS software improves interobserver agreement and accuracy of interpretation of aspects score. Interv. Neuroradiol. 2021, 27, 781–787. [Google Scholar] [CrossRef] [PubMed]
  8. Zhou, Z.; Rahman Siddiquee, M.M.; Tajbakhsh, N.; Liang, J. Unet++: A nested u-net architecture for medical image segmentation. In Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support: 4th International Workshop, DLMIA 2018, and 8th International Workshop, ML-CDS 2018, Held in Conjunction with MICCAI 2018, Granada, Spain, 20 September 2018; Proceedings 4; Springer International Publishing: Berlin/Heidelberg, Germany, 2018; pp. 3–11. [Google Scholar]
  9. Jha, D.; Smedsrud, P.H.; Riegler, M.A.; Johansen, D.; De Lange, T.; Halvorsen, P.; Johansen, H.D. Resunet++: An advanced architecture for medical image segmentation. In Proceedings of the 2019 IEEE International Symposium on Multimedia (ISM), San Diego, CA, USA, 9–11 December 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 225–2255. [Google Scholar]
  10. Conze, P.H.; Andrade-Miranda, G.; Singh, V.K.; Jaouen, V.; Visvikis, D. Current and emerging trends in medical image segmentation with deep learning. IEEE Trans. Radiat. Plasma Med. Sci. 2023, 7, 545–569. [Google Scholar] [CrossRef]
  11. Frid-Adar, M.; Diamant, I.; Klang, E.; Amitai, M.; Goldberger, J.; Greenspan, H. GAN-based synthetic medical image augmentation for increased CNN performance in liver lesion classification. Neurocomputing 2018, 321, 321–331. [Google Scholar] [CrossRef]
  12. Armanious, K.; Jiang, C.; Fischer, M.; Küstner, T.; Hepp, T.; Nikolaou, K.; Gatidis, S.; Yang, B. MedGAN: Medical image translation using GANs. Comput. Med. Imaging Graph. 2020, 79, 101684. [Google Scholar] [CrossRef]
  13. Tang, M.; Zhang, Z.; Cobzas, D.; Jagersand, M.; Jaremko, J.L. Segmentation-by-detection: A cascade network for volumetric medical image segmentation. In Proceedings of the 2018 IEEE 15th International Symposium on Biomedical Imaging (ISBI 2018), Washington, DC, USA, 4–7 April 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 1356–1359. [Google Scholar]
  14. Dou, Q.; Yu, L.; Chen, H.; Jin, Y.; Yang, X.; Qin, J.; Heng, P.A. 3D deeply supervised network for automated segmentation of volumetric medical images. Med. Image Anal. 2017, 41, 40–54. [Google Scholar] [CrossRef]
  15. Alarcón-Paredes, A.; Guzmán-Guzmán, I.P.; Hernández-Rosales, D.E.; Navarro-Zarza, J.E.; Cantillo-Negrete, J.; Cuevas-Valencia, R.E.; Alonso, G.A. Computer-aided diagnosis based on hand thermal, RGB images, and grip force using artificial intelligence as screening tool for rheumatoid arthritis in women. Med. Biol. Eng. Comput. 2021, 59, 287–300. [Google Scholar] [CrossRef]
  16. Mate, G.S.; Kureshi, A.K.; Singh, B.K. An efficient CNN for hand X-ray classification of rheumatoid arthritis. J. Healthc. Eng. 2021, 2021, 6712785. [Google Scholar] [CrossRef] [PubMed]
  17. Üreten, K.; Erbay, H.; Maraş, H.H. Detection of rheumatoid arthritis from hand radiographs using a convolutional neural network. Clin. Rheumatol. 2020, 39, 969–974. [Google Scholar] [CrossRef] [PubMed]
  18. Reed, M.; Le Souëf, T.; Rampono, E. Pilot study of a machine-learning tool to assist in the diagnosis of hand arthritis. Intern. Med. J. 2022, 52, 959–967. [Google Scholar] [CrossRef] [PubMed]
  19. Tkachenko, M.; Malyuk, M.; Holmanyuk, A.; Liubimov, N. Label Studio: Data Labeling Software. 2022. Available online: https://github.com/heartexlabs/label-studio (accessed on 16 February 2023).
  20. Isensee, F.; Jaeger, P.F.; Kohl, S.A.; Petersen, J.; Maier-Hein, K.H. nnU-Net: A self-configuring method for deep learning-based biomedical image segmentation. Nat. Methods 2021, 18, 203–211. [Google Scholar] [CrossRef] [PubMed]
  21. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015: 18th International Conference, Munich, Germany, 5–9 October 2015; Proceedings, Part III 18; Springer International Publishing: Berlin/Heidelberg, Germany, 2015; pp. 234–241. [Google Scholar]
  22. Ostmeier, S.; Axelrod, B.; Bertels, J.; Isensee, F.; Lansberg, M.G.; Christensen, S.; Albers, G.W.; Li, L.J.; Heit, J.J. Evaluation of Medical Image Segmentation Models for Uncertain, Small or Empty Reference Annotations. arXiv 2022, arXiv:2209.13008. [Google Scholar]
  23. Cardoso, M.J.; Li, W.; Brown, R.; Ma, N.; Kerfoot, E.; Wang, Y.; Murrey, B.; Myronenko, A.; Zhao, C.; Yang, D.; et al. Monai: An open-source framework for deep learning in healthcare. arXiv 2022, arXiv:2211.02701. [Google Scholar]
  24. Andersen, J.K.H.; Pedersen, J.S.; Laursen, M.S.; Holtz, K.; Grauslund, J.; Savarimuthu, T.R.; Just, S.A. Neural networks for automatic scoring of arthritis disease activity on ultrasound images. RMD Open 2019, 5, e000891. [Google Scholar] [CrossRef]
  25. Tripoliti, E.E.; Fotiadis, D.I.; Argyropoulou, M. Automated segmentation and quantification of inflammatory tissue of the hand in rheumatoid arthritis patients using magnetic resonance imaging data. Artif. Intell. Med. 2007, 40, 65–85. [Google Scholar] [CrossRef]
  26. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal loss for dense object detection. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar]
  27. Salehi, S.S.M.; Erdogmus, D.; Gholipour, A. Tversky loss function for image segmentation using 3D fully convolutional deep networks. In International Workshop on Machine Learning in Medical Imaging; Springer International Publishing: Cham, Switzerland, 2017; pp. 379–387. [Google Scholar]
  28. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial networks. Commun. ACM 2020, 63, 139–144. [Google Scholar] [CrossRef]
  29. Ho, J.; Jain, A.; Abbeel, P. Denoising diffusion probabilistic models. Adv. Neural Inf. Process. Syst. 2020, 33, 6840–6851. [Google Scholar]
  30. Fetty, L.; Bylund, M.; Kuess, P.; Heilemann, G.; Nyholm, T.; Georg, D.; Löfstedt, T. Latent space manipulation for high-resolution medical image synthesis via the StyleGAN. Z. Med. Phys. 2020, 30, 305–314. [Google Scholar] [CrossRef]
  31. Zhao, H.; Li, H.; Cheng, L. Synthesizing filamentary structured images with GANs. arXiv 2017, arXiv:1706.02185. [Google Scholar]
  32. Guan, Q.; Chen, Y.; Wei, Z.; Heidari, A.A.; Hu, H.; Yang, X.H.; Zheng, J.; Zhou, Q.; Chen, H.; Chen, F. Medical image augmentation for lesion detection using a texture-constrained multichannel progressive GAN. Comput. Biol. Med. 2022, 145, 105444. [Google Scholar] [CrossRef]
  33. Qasim, A.B.; Ezhov, I.; Shit, S.; Schoppe, O.; Paetzold, J.C.; Sekuboyina, A.; Kofler, F.; Lipkova, J.; Li, H.; Menze, B. Red-GAN: Attacking class imbalance via conditioned generation. Yet another medical imaging perspective. In Proceedings of the Third Conference on Medical Imaging with Deep Learning, PMLR, Montreal, QC, Canada, 6–8 July 2020; pp. 655–668. [Google Scholar]
Figure 1. 2 Example planar ⁹⁹mTc-maraciclatide scans of the hands showing focal joint inflammation with some inflamed regions magnified. (a) shows an oblique view, and (b) shows a palmar view.
Figure 1. 2 Example planar ⁹⁹mTc-maraciclatide scans of the hands showing focal joint inflammation with some inflamed regions magnified. (a) shows an oblique view, and (b) shows a palmar view.
Diagnostics 13 03298 g001
Figure 2. Example scans with clinically segmented labels of normal, low, and high inflammation of the tissue. Labels were segmented nonexclusively (each pixel can be labeled multiple times), and the label with the highest degree of inflammation was taken per pixel to force exclusivity. More information on this can be found in Section 2.1.
Figure 2. Example scans with clinically segmented labels of normal, low, and high inflammation of the tissue. Labels were segmented nonexclusively (each pixel can be labeled multiple times), and the label with the highest degree of inflammation was taken per pixel to force exclusivity. More information on this can be found in Section 2.1.
Diagnostics 13 03298 g002
Figure 3. Overview of training and testing of each model (threshold and nnU-Net described later in Section 2.2 and Section 2.3, respectively) for a single crossfold of data. The scans were separated into 90% train and 10% test; the two models were trained on the 90%. Once trained, the models were tested on the held-out 10% test dataset.
Figure 3. Overview of training and testing of each model (threshold and nnU-Net described later in Section 2.2 and Section 2.3, respectively) for a single crossfold of data. The scans were separated into 90% train and 10% test; the two models were trained on the 90%. Once trained, the models were tested on the held-out 10% test dataset.
Diagnostics 13 03298 g003
Figure 4. Thresholding model. The input image is first blurred by a single Gaussian kernel and then segmented using a lower and upper threshold for each of the three possible labels, resulting in 9 trainable parameters for the model. Each label is learned independently, and therefore, a pixel can be detected as multiple classes.
Figure 4. Thresholding model. The input image is first blurred by a single Gaussian kernel and then segmented using a lower and upper threshold for each of the three possible labels, resulting in 9 trainable parameters for the model. Each label is learned independently, and therefore, a pixel can be detected as multiple classes.
Diagnostics 13 03298 g004
Figure 5. Cross-folding diagram. Each model is trained and evaluated 10 times by using a different separate 10% of the dataset to evaluate the model and the other 90% to train the model.
Figure 5. Cross-folding diagram. Each model is trained and evaluated 10 times by using a different separate 10% of the dataset to evaluate the model and the other 90% to train the model.
Diagnostics 13 03298 g005
Figure 6. Low and high inflammation classes can be combined after model segmentation to create a single inflammation class. This can then be compared to the combined inflammation class created by combining the low and high classes segmented by the clinician. The combination is performed using pixel-wise addition of the segmentation maps. Thresholding model segmentations are clipped.
Figure 6. Low and high inflammation classes can be combined after model segmentation to create a single inflammation class. This can then be compared to the combined inflammation class created by combining the low and high classes segmented by the clinician. The combination is performed using pixel-wise addition of the segmentation maps. Thresholding model segmentations are clipped.
Diagnostics 13 03298 g006
Figure 7. Modified Dice score performance of the thresholding and nnU-Net models for the three classes as well as for all classes combined. The box-whisker plots represent the distribution in modified Dice score values based on the 10 models obtained from the 10 crossfolds. The boxes span from the lower to upper quartiles, showing the median in red. The whiskers extend to a maximum of 1.5 times the interquartile range but end at the lower/largest data point within the range; outliers beyond this range are plotted as circles.
Figure 7. Modified Dice score performance of the thresholding and nnU-Net models for the three classes as well as for all classes combined. The box-whisker plots represent the distribution in modified Dice score values based on the 10 models obtained from the 10 crossfolds. The boxes span from the lower to upper quartiles, showing the median in red. The whiskers extend to a maximum of 1.5 times the interquartile range but end at the lower/largest data point within the range; outliers beyond this range are plotted as circles.
Diagnostics 13 03298 g007
Figure 8. Normalized pixel confusion matrices for the two models and the human intraobserver measurements. Confusion matrices were plotted with modified code from http://scikit-learn.org/stable/auto_examples/model_selection/plot_confusion_matrix.html (accessed on 28 August 2023).
Figure 8. Normalized pixel confusion matrices for the two models and the human intraobserver measurements. Confusion matrices were plotted with modified code from http://scikit-learn.org/stable/auto_examples/model_selection/plot_confusion_matrix.html (accessed on 28 August 2023).
Diagnostics 13 03298 g008
Figure 9. Two example sets of generated labels for Thresholding and nnU-Net models.
Figure 9. Two example sets of generated labels for Thresholding and nnU-Net models.
Diagnostics 13 03298 g009
Figure 10. ROC curve calculated by varying the threshold number of pixels in the prediction to classify the image as inflamed.
Figure 10. ROC curve calculated by varying the threshold number of pixels in the prediction to classify the image as inflamed.
Diagnostics 13 03298 g010
Table 1. Modified Dice score segmentation results by crossfolding over the held-out test dataset. Best highlighted in bold.
Table 1. Modified Dice score segmentation results by crossfolding over the held-out test dataset. Best highlighted in bold.
CrossfoldThresholding ModelnnU-Net Model
NormalLowHighAllNormalLowHighAll
10.930.20.280.470.950.610.720.76
20.930.080.490.50.940.380.80.71
30.910.120.320.450.940.430.650.67
40.910.190.110.40.950.460.440.61
50.920.050.330.430.940.640.930.83
60.920.110.680.570.940.50.840.76
70.910.240.730.630.940.610.90.82
80.920.250.130.440.930.290.640.62
90.930.110.170.40.940.750.960.88
100.900.080.30.430.930.430.730.7
Mean0.92 ± 0.010.14 ± 0.070.35 ± 0.210.47 ± 0.080.94 ± 0.010.51 ± 0.140.76 ± 0.160.74 ± 0.09
Table 2. Modified Dice score segmentation results by crossfolding over the held-out test dataset when collapsing low and high inflammation into a single class. Best highlighted in bold.
Table 2. Modified Dice score segmentation results by crossfolding over the held-out test dataset when collapsing low and high inflammation into a single class. Best highlighted in bold.
CrossfoldThresholding ModelnnU-Net Model
10.340.73
20.50.68
30.390.67
40.20.66
50.260.83
60.680.82
70.720.83
80.280.5
90.210.89
100.420.63
Mean0.4 ± 0.180.72 ± 0.12
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Cobb, R.; Cook, G.J.R.; Reader, A.J. Deep Learned Segmentations of Inflammation for Novel ⁹⁹mTc-maraciclatide Imaging of Rheumatoid Arthritis. Diagnostics 2023, 13, 3298. https://doi.org/10.3390/diagnostics13213298

AMA Style

Cobb R, Cook GJR, Reader AJ. Deep Learned Segmentations of Inflammation for Novel ⁹⁹mTc-maraciclatide Imaging of Rheumatoid Arthritis. Diagnostics. 2023; 13(21):3298. https://doi.org/10.3390/diagnostics13213298

Chicago/Turabian Style

Cobb, Robert, Gary J. R. Cook, and Andrew J. Reader. 2023. "Deep Learned Segmentations of Inflammation for Novel ⁹⁹mTc-maraciclatide Imaging of Rheumatoid Arthritis" Diagnostics 13, no. 21: 3298. https://doi.org/10.3390/diagnostics13213298

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop