Next Article in Journal
Single-Center Experience with Swiss LithoClast® Trilogy for Kidney Stones
Previous Article in Journal
Breast Sarcomas—How Different Are They from Breast Carcinomas? Clinical, Pathological, Imaging and Treatment Insights
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Specific Binding Ratio Estimation of [123I]-FP-CIT SPECT Using Frontal Projection Image and Machine Learning

1
Radiological Center, University of Fukui Hospital, 23-3, Matsuoka-Shimoaizuki, Eiheiji-cho, Fukui 910-1193, Japan
2
Biomedical Imaging Research Center, University of Fukui, 23-3, Matsuoka-Shimaizuki, Eiheiji-cho, Fukui 910-1193, Japan
3
Department of Radiology, Faculty of Medical Sciences, University of Fukui, 23-3, Matsuoka-Shimoaizuki, Eiheiji-cho, Fukui 910-1193, Japan
*
Authors to whom correspondence should be addressed.
Diagnostics 2023, 13(8), 1371; https://doi.org/10.3390/diagnostics13081371
Submission received: 10 February 2023 / Revised: 3 April 2023 / Accepted: 6 April 2023 / Published: 7 April 2023
(This article belongs to the Section Medical Imaging and Theranostics)

Abstract

:
This study aimed to develop a new convolutional neural network (CNN) method for estimating the specific binding ratio (SBR) from only frontal projection images in single-photon emission-computed tomography using [123I]ioflupane. We created five datasets to train two CNNs, LeNet and AlexNet: (1) 128FOV used a 0° projection image without preprocessing, (2) 40FOV used 0° projection images cropped to 40 × 40 pixels centered on the striatum, (3) 40FOV training data doubled by data augmentation (40FOV_DA, left-right reversal only), (4) 40FOVhalf, and (5) 40FOV_DAhalf, split into left and right (20 × 40) images of 40FOV and 40FOV_DA to separately evaluate the left and right SBR. The accuracy of the SBR estimation was assessed using the mean absolute error, root mean squared error, correlation coefficient, and slope. The 128FOV dataset had significantly larger absolute errors compared to all other datasets (p < 0. 05). The best correlation coefficient between the SBRs using SPECT images and those estimated from frontal projection images alone was 0.87. Clinical use of the new CNN method in this study was feasible for estimating the SBR with a small error rate using only the frontal projection images collected in a short time.

1. Introduction

Single-photon emission-computed tomography (SPECT) can show the distribution of the dopamine transporter (DAT) in the brain using [123I] ioflupane ([123I]-FP-CIT), the DAT imaging ligand. Employing [123I]-FP-CIT SPECT increases the precision of choosing the best treatment options and aids in the early detection of Parkinsonian syndrome and dementia with Lewy bodies [1,2,3]. Visual evaluation is reportedly sufficient for diagnosis [1,4]; however, diagnostic performance is enhanced by quantitative evaluation in addition to visual inspection [5,6,7]. The specific binding ratio (SBR), calculated as the ratio between specific and non-specific accumulation, is the standard quantification parameter in [123I]-FP-CIT SPECT. Currently, in clinical practice, quantitative evaluation by SBR plays an important role as an adjunct to visual evaluation.
Worldwide, [123I]-FP-CIT SPECT exams are often conducted, and clinical practice guidelines have been published in various regions [8,9]. [123I]-FP-CIT SPECT scans in Japan are performed in accordance with the ‘Clinical Ioflupane Guidelines’ of the Japanese Society of Nuclear Medicine, available online at http://www.jsnm.org/archives/1151/ (accessed on 6 April 2023). To get sufficient acquisition counts, these guidelines indicate an appropriate acquisition time of 30 to 45 min. Short-time scans reduce the acquisition counts, deteriorate the quality of the visual evaluation, and underestimate SBR [10]. In addition, a multicenter study conducted in Japan revealed that all 14 participating facilities used an acquisition time of approximately 30 min [11], suggesting that many hospitals in Japan comply with the guidelines. However, patients with claustrophobia, back pain, dementia, etc., often request shorter scans or want to stop immediately after the scan is started. If the examination is terminated without sufficient scanning time, a tomographic image may not be obtained, and the SPECT study would fail to obtain SBR and any other information. Although [123I]-FP-CIT SPECT is important to evaluate striatal functions, some patients cannot tolerate the static position for reasons described above. If a frontal projection image with only a short time duration can estimate striatal functions in such cases, the image analysis method may provide appropriate clinical diagnosis.
Our study aimed to develop a new CNN method for estimating SBR from only frontal projection images in [123I]-FP-CIT SPECT, and it also aimed to evaluate the error and correlation between SBR estimated by the new CNN method and SBR obtained by the full SPECT study. If patients who cannot undergo the exam for the required amount of time could receive an accurate estimate of SBR from a rapid frontal projection, the results of our investigation would be of great clinical value. Nakajima et al. developed a machine learning (ML) method that could diagnose Parkinsonian syndrome and dementia with Lewy bodies [12]. They obtained more accurate results than conventional ROI-based methods. However, the purpose of their study was to improve diagnostic accuracy using commonly obtained SPECT images, and the concept was different from our study.

2. Materials and Methods

2.1. SPECT Imaging

Four hundred twenty patients (72.1 ± 9.7 y) who underwent [123I]-FP-CIT SPECT at the University of Fukui Hospital from January 2017 to December 2021 were included in this retrospective study. The Ethics Committee of our university’s Faculty of Medical Sciences approved the study’s protocol (#20210156) and waived the requirement of obtaining patient informed consent and approved a retrospective evaluation of imaging and clinical data. Patients received a gradual intravenous infusion of approximately 170 MBq [123I]-FP-CIT in the morning, and 3–4 h later, they had SPECT scans (167 MBq at noon).
A dual-head gamma camera (Symbia T2, Siemens, Erlargen, Germany) and low- to medium-energy general-purpose collimators were installed in the SPECT/CT scanner. The SPECT images were taken with a 1.45 zoom, a 3.3 mm pixel size, and a 10% scan energy window in close proximity to the 159-keV photo-peak. With the triple-energy window scatter correction (SC), two sub-windows at 7% were constructed on the top and bottom sides of the photo-peak window [13]. With an angular range of 4° increments, 90 frames were used to get the SPECT data. The camera heads were as near to the patient’s head as possible. An attenuation correction low-dose CT scan was performed following the gathering of SPECT data. The data were acquired in continuous rotation mode at 14 min × two repeats. With eight iterations and ten subsets, the SPECT data were reconstructed using an iterative three-dimensional-ordered subset expectation maximization algorithm. After 1.5× expansion, the pixel size of each slice in a 128 × 128 matrix was 2.2 mm, and all images were subjected to a Gaussian filter with a full width at half maximum of 6 mm. Two image sets were reconstructed from the image data with the correction of CT attenuation and SC (ACSC) and without any correction (non-AC).

2.2. SBR Calculation

The SBR values were calculated with DaT View (AZE, Tokyo, Japan) using the Tossici-Bolt and count-based automatic calculation methods written as a Matlab script [14,15,16]. DaT View is a widely used SBR calculation software in Japan that recommends using non-AC images as input images [17,18]. In the count-based method (the most commonly used calculation method in the world), SBR is calculated from region of interest (ROI) counts based on the formula (striatum ROI—BG)/BG, where BG is background counts. Our semi-automatic calculation method used in this study can be applied for either ACSC or CTAC images, and ACSC is recommended [15,16]. Thus, we applied non-AC images for DaT View, and we applied ACSC images for the count-based method.

2.3. Network Architecture

Only the front (0°) projection image from the projection image acquired by SPECT was defined as the 0° projection image. The new CNN method for SBR estimation was constructed using CNN upon studying the combination of the 0° projection image and the SBRs from the full SPECT data.
The 0° projection image was a total image of 4 degrees (0° to 3°) of the angle at the time of SPECT continuous rotation acquisition, which was equivalent to 37.3 s of acquisition time. Two types of datasets were constructed: one for estimating the left and right average SBR values, and the other for estimating the left and right SBRs independently. The Neural Network (v2. 10, Sony Network Communications Inc., Tokyo, Japan) was used for deep learning (DL) in our CNN models for SBR estimation from the 0° projection image. Input (I), rectified linear unit (R, ReLU), convolution (C), max-pooling (M), affine (A), Tanh (T), and squared error (S) made up the architecture used in this investigation (Figure S1 in Supplementary Materials). The layer’s function is indicated by square boxes, and the specifications for each layer are depicted in the figures on the right side of the box. The three values on the right side of the first input layer, for example, denote the number of colors (“1” signifies monochrome) and the size of the input image (height and width). In the second convolution layer, the output feature map’s number and size (height and width) were both displayed in the same format. The rectified linear unit is referred to as “ReLU”. The home page describes additional details in this figure (Sony Network Communications Inc., 2020, available online at https://dl.sony.com/ (accessed on 6 April 2023)). We tested two types of networks, LeNet [19] (Figure 1a) and AlexNet [20] (Figure 1b), based on CNN.
CNN is one of the commonly used Deep Learning architecture types for identifying and classifying images. It has several layers, including fully connected, convolutional, and pooling layers. The convolutional layer uses filters to extract features from the input image, the pooling layer minimizes processing by downsampling the image, and the fully connected layer provides the final prediction. Innovative CNN topologies such as LeNet and AlexNet were developed to increase accuracy and lower computing costs. LeNet, also known as the classic neural network, was developed by LeCun et al. in the 1990s for the recognition of handwritten and machine-printed characters [19]. The architecture is relatively uncomplicated and easy to comprehend. However, AlexNet has more layers and more filters than LeNet, making the network more complex and more accurate.

2.4. Datasets

Five datasets were created. Figure 2 displays the images utilized in each dataset and the quantity of training, validation, and test cases. The 128FOV dataset consisted of a combination of 260, 80, and 80 cases of training, validation, and test data, respectively (Figure 2a), and used a 0° projection image without preprocessing. Table 1 shows the characteristics of the patients included in the datasets. No significant differences were observed with respect to their age and gender. They were suspected to have Parkinsonian syndrome or dementia with Lewy bodies. The dataset 40FOV used 0° projection images cropped to 40 × 40 pixels centered on the striatum (the number of testing, validation, and training cases was the same as for 128FOV) (Figure 2b). The dataset 40FOV_DA was the 40FOV training data doubled by data augmentation (left-right reversal only), with 520, 80, and 80 cases of training, validation, and test data, respectively (Figure 2c). In addition, for the 40FOV and 40FOV_DA, the 40 × 40 pixels cropped image was split into left and right (20 × 40) so that the left and right SBR could be evaluated separately. We named those datasets 40FOVhalf and 40FOV_DAhalf, respectively (Figure 2d,e).

2.5. Evolution

The SBR obtained by DaT View and the count-based method were defined as SBRTB and SBRcb, respectively. The SBR estimated by DL using 0° projection images was defined as SBRDL. Using the root mean square error (RMSE), correlation coefficient (Corr) mean absolute error (MAE), and slope, we assessed the accuracy of the SBR estimate. In addition, the CNN-focused regions of the 0° projection image were identified by Gradient-weighted Class Activation Mapping (Grad-CAM) [21]. According to their significance, Grad-CAM created a heat map of the pixels that CNN concentrated on.

2.6. Statistical Analysis

To compare absolute errors between datasets, an analysis of variance (ANOVA) was performed, followed by Tukey’s test for post hoc testing. To compare the MAE, RSMU, and Corr across the AlexNet and LeNet networks, the Wilcoxon signed-rank test was utilized. EZR ver. 1.41 [22] was used for the statistical analysis, and p < 0.05 was regarded as significant.

3. Results

Figure 3 shows the scatter plots of the conventional SBR and the SBR for each dataset and network, and the corresponding Bland–Altman plot for each SBR. SBRBT vs. SBRDL is shown in Figure 3a, and SBRcb vs. SBRDL is shown in Figure 3b. The solid line is the line of identity. The Bland–Altman plot shows that 128FOV has a distribution with a biased plot compared to 40FOV and 40FOV_DA. The trend was similar for both the DaT View and the count-based methods. Table 2 shows the values of the MAE, RSMU, Corr, and slope obtained from the DaT View and count-based method. The 128FOV had larger MAEs and RMSEs and smaller Corrs and slopes compared to the other datasets. The 128FOV had significantly larger MAEs compared to the other datasets (p < 0.05). Apart from the 128FOV combination, there was no significant variation in MAE and RMSE, and the values did not differ considerably between AlexNet and LeNet. Data augmentation had slightly improved estimates. SBRBT and SBRcb showed similar trends.
Figure 4 shows the original 0° projection image and the heat map image of Grad-CAM for 128FOV (Figure 4a, top) and 40FOV (Figure 4a, bottom). In the Grad-CAM of 128FOV, the many background dots indicate that the CNN focused on the background and the brain in many cases. The 40FOV of the same case at the bottom shows that the small FOV was centered on the striatum, and the CNN focused on the brain. Figure 4b includes [123I]-FP-CIT uptake in the parotid gland. In the Grad-CAM of the 128FOV (Figure 4b, top), the color dots in the parotid gland indicate that the CNN focused on the parotid gland in addition to the background. On the other hand, the image of the 40FOV (Figure 4b, bottom) was cropped around the striatum, and the parotid gland was out of the field of view, indicating that the CNN was not focused on parotid accumulation.

4. Discussion

The purpose of this study was to create a new CNN method for estimating SBR from only frontal projection images in [123I]-FP-CIT SPECT using CNN and to evaluate the error and correlation between the SBR estimated by the new CNN method and the full SPECT study. The new CNN method estimated the SBR with high correlation and low error using the preprocessed dataset for the projection image. In addition, although the dataset for evaluating the left and right SBR separately (40FOVhalf and 40FOV_DAhalf) had larger errors than the 40FOV and 40FOV_DA, it demonstrated a strong correlation between the SBRs with full SPECT data and the SBRDL. Bland–Altman plots revealed that the 128FOV dataset had a larger systematic error than the 40FOV and 40FOV_DA datasets. This suggests that the increase in cases did not improve the SBR estimation accuracy in the 128FOV dataset. Therefore, 128FOV projection data alone (image without trimming processing) are inappropriate for SBR estimation. It seems that noise in the large space outside the head may have affected this result because 40FOV shows a slightly similar trend. These results showed the same tendency in both the DaT View and the count-based methods. We consider calculating the left and right SBR independently to be clinically valuable since patients with Parkinsonian syndrome often have left–right differences in SBR [23]. Furthermore, the number of data is doubled by dividing the data into left and right, which improves the accuracy of DL. Since this study yielded low error rates, the results may not be equivalent to full SPECT data (Figure 3). Therefore, our method may be used as an alternative to SBR only when collected for a short scanning time.
The results using the 128FOV were significantly poorer than the other datasets. The first possible reason for this could be that the CNN often focused equally on the background and on the brain (Figure 4a). This was probably because the area occupied by the background was larger than the area occupied by the brain in the images of the 128FOV. Although the background signal was not involved in the SBR calculation, the background noise was assumed to induce confusion in the learning process. Furthermore, in cases with accumulation in the salivary glands (Figure 4b), the accumulation could affect the learning process with 128FOV. Physiological accumulation of the salivary glands should be excluded from the calculation of SBR [24]. Therefore, the projection data without trimming are considered inappropriate for SBR estimation. The results of the estimation using data augmentation approaches were marginally better than those without data augmentation techniques, indicating that the accuracy of the SBR estimation was enhanced by the increase in the number of cases. The accuracy of the SBR estimation would be enhanced by additional data augmentation, which would increase the number of cases and training data.
There was no significant difference between the two networks. Therefore, we believe that by optimizing these networks and testing networks with more layers, such as VGGNet [25] and ResNet [26], we could find the optimal network for SBR estimation; moreover, further improvement can be anticipated in terms of accurate SBR estimation.

5. Conclusions

The 0° projection image in [123I]-FP-CIT SPECT estimated the SBR with high correlation and low errors by using a preprocessed dataset for the projection image. The estimation accuracy was improved by increasing the training cases through data augmentation. Clinical use of the new CNN method in this study is considered feasible since SBR can be estimated with a small error rate using only frontal projection images collected in a short time.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/diagnostics13081371/s1, Figure S1: Two types of network diagrams, (a) LeNet and (b) AlexNet. The layer’s function is indicated by square boxes. The specifications for each layer are depicted in the figures on the right side of the box. The web site https://dl.sony.com/ (accessed on 6 April 2023) describes additional details in this figure (Sony Network Communications Inc., 2020).

Author Contributions

A.K. and H.O. contributed to the conception and design of the study. A.K., K.S. and E.K. contributed to acquiring and interpreting data. A.K. and H.O. contributed to the statistical analysis. A.K., H.O., N.K. and T.T. contributed to the analysis and interpretation of the data. All authors have read and agreed to the published version of the manuscript.

Funding

The Japan Society for the Promotion of Science (JSPS) KAKENHI Grant No. 22K15842 provided funding for this project.

Institutional Review Board Statement

This study was conducted according to the guidelines of the Declaration of Helsinki and approved by the Ethics Committee of Fukui University, Faculty of Medical Sciences (#20210156 and 11 January 2022).

Informed Consent Statement

Need for obtaining patient consent was waived due to the retrospective nature of the study.

Data Availability Statement

The datasets used and/or analyzed during the current study are available from the corresponding author upon reasonable request.

Acknowledgments

The facility staff members who contributed to this research are all gratefully acknowledged by the authors.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Catafau, A.M.; Tolosa, E.; DaTSCAN Clinically Uncertain Parkinsonian Syndromes Study Group. Impact of dopamine transporter SPECT using 123I-Ioflupane on diagnosis and management of patients with clinically uncertain Parkinsonian syndromes. Mov. Disord. 2004, 19, 1175–1182. [Google Scholar] [CrossRef] [PubMed]
  2. Walker, Z.; Jaros, E.; Walker, R.W.H.; Lee, L.; Costa, D.C.; Livingston, G.; Ince, P.G.; Perry, R.; Mckeith, I.; Katona, C.L.E. Dementia with Lewy bodies: A comparison of clinical diagnosis, FP-CIT single photon emission computed tomography imaging and autopsy. J. Neurol. Neurosurg. Psychiatry 2007, 78, 1176–1181. [Google Scholar] [CrossRef] [PubMed]
  3. McKeith, I.; O’Brien, J.; Walker, Z.; Tatsch, K.; Booij, J.; Darcourt, J.; Padovani, A.; Giubbini, R.; Bonuccelli, U.; Volterrani, D.; et al. Sensitivity and specificity of dopamine transporter imaging with 123I-FP-CIT SPECT in dementia with Lewy bodies: A phase III, multicentre study. Lancet Neurol. 2007, 6, 305–313. [Google Scholar] [CrossRef] [PubMed]
  4. Benamer, T.S.; Patterson, J.; Grosset, D.G.; Booij, J.; de Bruin, K.; van Royen, E.; Speelman, J.D.; Horstink, M.H.I.M.; Sips, H.J.W.A.; Dierckx, R.A.; et al. Accurate differentiation of parkinsonism and essential tremor using visual assessment of [123I]FP-CIT SPECT imaging. Mov. Disord. 2000, 15, 503–510. [Google Scholar] [CrossRef] [PubMed]
  5. Badiavas, K.; Molyvda, E.; Iakovou, I.; Tsolaki, M.; Psarrakos, K.; Karatzas, N. SPECT imaging evaluation in movement disorders: Far beyond visual assessment. Eur. J. Nucl. Med. Mol. Imaging 2011, 38, 764–773. [Google Scholar] [CrossRef]
  6. Booij, J.; Habraken, J.B.; Bergmans, P.; Tissingh, G.; Winogrodzka, A.; Wolters, E.C.; Janssen, A.G.; Stoof, J.C.; van Royen, E.A. Imaging of dopamine transporters with iodine-123-FP-CIT SPECT in healthy controls and patients with Parkinson’s disease. J. Nucl. Med. 1998, 39, 1879–1884. [Google Scholar]
  7. Soderlund, T.A.; Dickson, J.C.; Prvulovich, E.; Ben-Haim, S.; Kemp, P.; Booij, J.; Nobili, F.; Thomsen, G.; Sabri, O.; Koulibaly, P.M.; et al. Value of semiquantitative analysis for clinical reporting of 123I-2-β-carbomethoxy-3β-(4-iodophenyl)-N-(3-fluoropropyl) nortropane SPECT studies. J. Nucl. Med. 2013, 54, 714–722. [Google Scholar] [CrossRef] [Green Version]
  8. Djang, D.S.; Janssen, M.J.; Bohnen, N.; Booij, J.; Henderson, T.A.; Herholz, K.; Minoshima, S.; Rowe, C.C.; Sabri, O.; Seibyl, J.; et al. SNM Practice Guideline for Dopamine Transporter Imaging with I-123-Ioflupane SPECT 1. 0. J. Nucl. Med. 2012, 53, 154–163. [Google Scholar] [CrossRef] [Green Version]
  9. Darcourt, J.; Booij, J.; Tatsch, K.; Varrone, A.; Vander Borght, T.; Kapucu, O.L.; Nagren, K.; Nobili, F.; Walker, Z.; Van Laere, K. EANM procedure guidelines for brain neurotransmission SPECT using 123I-labelled dopamine transporter ligands, version 2. Eur. J. Nucl. Med. Mol. Imaging 2010, 37, 443–450. [Google Scholar] [CrossRef]
  10. Kita, A.; Okazawa, H.; Sugimoto, K.; Kaido, R.; Kidoya, E.; Kimura, H. Acquisition count dependence of the specific binding ratio in 123I-FP-CIT SPECT. Ann. Nucl. Med. 2021, 62, 3028. [Google Scholar] [CrossRef]
  11. Kita, A.; Onoguchi, M.; Shibutani, T.; Horita, H.; Oku, Y.; Kashiwaya, S.; Isaka, M.; Saitou, M. Standardization of the specific binding ratio in [123I]FP-CIT SPECT: Study by striatum phantom. Nucl. Med. Commun. 2019, 40, 484–490. [Google Scholar] [CrossRef]
  12. Nakajima, K.; Saito, S.; Chen, Z.; Komatsu, J.; Maruyama, K.; Shirasaki, N.; Watanabe, S.; Inaki, A.; Ono, K.; Kinuya, S. Diagnosis of Parkinson syndrome and Lewy-body disease using 123I-ioflupane images and a model with image features based on machine learning. Ann. Nucl. Med. 2022, 36, 765–776. [Google Scholar] [CrossRef]
  13. Ogawa, K.; Harata, Y.; Ichihara, T.; Kubo, A.; Hashimoto, S.; Kashiwaya, S. A practical method for position-dependent Compton-scatter correction in single photon emission CT. IEEE Trans. Med. Imaging 1991, 10, 408–412. [Google Scholar] [CrossRef]
  14. Tossici-Bolt, L.; Hoffmann, S.M.; Kemp, P.M.; Mehta, R.L.; Fleming, J.S. Quantification of [123I]FP-CIT SPECT brain images: An accurate technique measurement of the specific binding ratio. Eur. J. Nucl. Med. Mol. Imaging 2006, 33, 1491–1499. [Google Scholar] [CrossRef]
  15. Rahman, M.G.M.; Islam, M.M.; Tsujikawa, T.; Kiyono, Y.; Okazawa, H. Count-based method for specific binding ratio calculation in [I-123]FP-CIT SPECT analysis. Ann. Nucl. Med. 2019, 33, 14–21. [Google Scholar] [CrossRef] [Green Version]
  16. Rahman, M.G.M.; Islam, M.M.; Tsujikawa, T.; Okazawa, H. A novel automatic approach for calculation of the specific binding ratio in [i-123]fp-citspect. Diagnostics 2020, 10, 289. [Google Scholar] [CrossRef]
  17. Iwabuchi, Y.; Nakahara, T.; Kameyama, M.; Yamada, Y.; Hashimoto, M.; Matsusaka, Y.; Osada, T.; Ito, D.; Tabuchi, H.; Jinzaki, M. Impact of a combination of quantitative indices representing uptake intensity, shape, and asymmetry in DAT SPECT using machine learning: Comparison of different volume of interest settings. EJNMMI Res. 2019, 9, 7. [Google Scholar] [CrossRef]
  18. Murakami, H.; Kimura, A.; Yasumoto, T.; Miki, A.; Yamamoto, K.; Ito, N.; Momma, Y.; Owan, Y.; Yano, S.; Ono, K. Usefulness differs between the visual assessment and specific binding ratio of 123i-ioflupane spect in assessing clinical symptoms of drug-naïve parkinson’s disease patients. Front. Aging Neurosci. 2018, 10, 412. [Google Scholar] [CrossRef]
  19. LeCun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef] [Green Version]
  20. Krizhevsky, A.; Sutskever, I.; Hinton, E.G. ImageNet classification with deep convolutional neural networks. In Proceedings of the NIPS’12 25th International Conference on Neural Information Processing Systems, New York, NY, USA, 3 December 2012; Volume 1, pp. 1097–1105. [Google Scholar]
  21. Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-cam: Visual explanations from deep networks via gradient-based localization. Int. J. Comput. Vis. 2020, 128, 336–359. [Google Scholar] [CrossRef] [Green Version]
  22. Kanda, Y. Investigation of the freely available easy-to-use software “EZR” for medical statistics. Bone Marrow Transpl. 2013, 48, 452–458. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  23. Shigekiyo, T.; Arawaka, S. Laterality of specific binding ratios on DAT-SPECT for differential diagnosis of degenerative parkinsonian syndromes. Sci. Rep. 2020, 10, 15761. [Google Scholar] [CrossRef] [PubMed]
  24. Booij, J.; Busemann Sokole, E.; Stabin, M.G.; Janssen, A.G.; de Bruin, K.; van Royen, E.A. Human biodistribution and dosimetry of [123I]FP-CIT: A potent radioligand for imaging of dopamine transporters. Eur. J. Nucl. Med. 1998, 25, 24–30. [Google Scholar] [CrossRef] [PubMed]
  25. Szegedy, C.; Wei, L.; Yangqing, J.; Pierre, S.; Scott, R.; Scott, R.; Dragomir, A.; Dumitru, E.; Vincent, V.; Andrew, R. Going deeper with convolutions. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar] [CrossRef] [Green Version]
  26. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2015; pp. 770–778. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Scheme of the algorithms of ReNet and Alexnet used in this study. The number under the square boxes is the number of filters. LeNet (a), the basic shapes of CNN, consists of alternating layers of convolutional and pooling layers, and AlexNet (b) consists of five convolutional layers and three fully connected layers.
Figure 1. Scheme of the algorithms of ReNet and Alexnet used in this study. The number under the square boxes is the number of filters. LeNet (a), the basic shapes of CNN, consists of alternating layers of convolutional and pooling layers, and AlexNet (b) consists of five convolutional layers and three fully connected layers.
Diagnostics 13 01371 g001
Figure 2. 128FOV: combination of 260, 80, and 80 cases of training, validation, and test data, respectively. (a) 128FOV used a 0° projection image without preprocessing. (b) 40FOV used 0° projection images cropped to 40 × 40 pixels centered on the striatum (the number of testing, validation, and training cases was the same as for 128FOV). (c) 40FOV_DA used 40FOV training data doubled by data augmentation (left-right reversal only), with 520, 80, and 80 cases of training, validation, and test data, respectively. (d) 40FOVhalf used the image from 40FOV, which was divided in half (20 × 40) at the center of the left and right striatum, with 520, 80, and 80 cases of training, validation, and test data, respectively. (e) 40FOV_DAhalf used the image from 40FOV_DA, which was divided in half (20 × 40) at the center of the left and right striatum, with 1040, 80, and 80 cases of training, validation, and test data, respectively.
Figure 2. 128FOV: combination of 260, 80, and 80 cases of training, validation, and test data, respectively. (a) 128FOV used a 0° projection image without preprocessing. (b) 40FOV used 0° projection images cropped to 40 × 40 pixels centered on the striatum (the number of testing, validation, and training cases was the same as for 128FOV). (c) 40FOV_DA used 40FOV training data doubled by data augmentation (left-right reversal only), with 520, 80, and 80 cases of training, validation, and test data, respectively. (d) 40FOVhalf used the image from 40FOV, which was divided in half (20 × 40) at the center of the left and right striatum, with 520, 80, and 80 cases of training, validation, and test data, respectively. (e) 40FOV_DAhalf used the image from 40FOV_DA, which was divided in half (20 × 40) at the center of the left and right striatum, with 1040, 80, and 80 cases of training, validation, and test data, respectively.
Diagnostics 13 01371 g002
Figure 3. Scatter plots of conventional specific binding rate (SBR) (SBRBT or SBRcb) and SBR (SBRDL) for each dataset and network using the DaT view (a, left half) and the count-based methods (b, left half). The solid line is the line of identity. Right half shows the corresponding Bland–Altman plot for each SBR. Black lines represent the mean bias; the upper and lower dotted lines are 1.96 standard deviation.
Figure 3. Scatter plots of conventional specific binding rate (SBR) (SBRBT or SBRcb) and SBR (SBRDL) for each dataset and network using the DaT view (a, left half) and the count-based methods (b, left half). The solid line is the line of identity. Right half shows the corresponding Bland–Altman plot for each SBR. Black lines represent the mean bias; the upper and lower dotted lines are 1.96 standard deviation.
Diagnostics 13 01371 g003aDiagnostics 13 01371 g003b
Figure 4. Upper part, image of the 128FOV and Gradient-weighted Class Activation Mapping (Grad-CAM)’s heat map; lower part, image of the 40FOV of the same case and the heat map image of Grad-CAM. (a) In the Grad-CAM of 128FOV, the many background dots indicate that the convolutional neural network (CNN) focused on the background. The 40FOV of the same case in the bottom shows that the small FOV was centered on the striatum, and CNN focused on the brain. (b) A case of [123I] ioflupane ([123I]-FP-CIT) accumulation in the parotid gland. In the Grad-CAM of the 128FOV (top), the color dots in the parotid gland indicate that the CNN focused on the parotid gland in addition to the background, while the image of the 40FOV (bottom) was cropped around the striatum, and the parotid gland was outside the field of view, indicating that the CNN was not interested in parotid accumulation.
Figure 4. Upper part, image of the 128FOV and Gradient-weighted Class Activation Mapping (Grad-CAM)’s heat map; lower part, image of the 40FOV of the same case and the heat map image of Grad-CAM. (a) In the Grad-CAM of 128FOV, the many background dots indicate that the convolutional neural network (CNN) focused on the background. The 40FOV of the same case in the bottom shows that the small FOV was centered on the striatum, and CNN focused on the brain. (b) A case of [123I] ioflupane ([123I]-FP-CIT) accumulation in the parotid gland. In the Grad-CAM of the 128FOV (top), the color dots in the parotid gland indicate that the CNN focused on the parotid gland in addition to the background, while the image of the 40FOV (bottom) was cropped around the striatum, and the parotid gland was outside the field of view, indicating that the CNN was not interested in parotid accumulation.
Diagnostics 13 01371 g004
Table 1. Patient characteristics.
Table 1. Patient characteristics.
TrainingValidationTest
Number of cases2608080
Age (y, mean ± SD)72.2 ± 9.671.8 ± 10.372.4 ± 9.8
Men/women (N)137/12349/3144/36
Table 2. Results of DaT View and count-based methods analysis. MAE, RSMU, Corr, and slope values for the datasets 128FOV, 40FOV, 40FOV_DA, 40FOVhalf, and 40FOV_DAhalf.
Table 2. Results of DaT View and count-based methods analysis. MAE, RSMU, Corr, and slope values for the datasets 128FOV, 40FOV, 40FOV_DA, 40FOVhalf, and 40FOV_DAhalf.
LeNetAlexNet
DatasetMAERMSECorrSlopeMAERMSECorrSlope
DaT View128FOV1.561.870.360.211.211.510.440.16
40FOV0.921.180.740.600.861.090.780.47
40FOV_DA0.831.050.800.740.760.940.840.74
40FOVhalf1.001.250.710.670.981.210.720 60
40FOV_DAhalf0.941.200.740.730.951.180.730.61
Count-based methods128FOV0.580.720.410.230.630.760.250.10
40FOV0.350.420.840.700.480.570.840.36
40FOV_DA0.350.440.830.760.350.420.870.59
40FOVhalf0.450.550.730.590.390.500.780.64
40FOV_DAhalf0.400.490.800.710.400.500.770.62
MAE: mean absolute error, RMSE: root mean squared error, Corr: correlation coefficient. 128FOV had a larger MAE and RMSE than the other datasets, and the Corr and slope was small; only weak correlations were obtained. When AlexNet was used for the network, there was little difference in the values when LeNet was used. Data augmentation had a slight effect, slightly improving estimates. SBRBT and SBRcb showed similar trends.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Kita, A.; Okazawa, H.; Sugimoto, K.; Kosaka, N.; Kidoya, E.; Tsujikawa, T. Specific Binding Ratio Estimation of [123I]-FP-CIT SPECT Using Frontal Projection Image and Machine Learning. Diagnostics 2023, 13, 1371. https://doi.org/10.3390/diagnostics13081371

AMA Style

Kita A, Okazawa H, Sugimoto K, Kosaka N, Kidoya E, Tsujikawa T. Specific Binding Ratio Estimation of [123I]-FP-CIT SPECT Using Frontal Projection Image and Machine Learning. Diagnostics. 2023; 13(8):1371. https://doi.org/10.3390/diagnostics13081371

Chicago/Turabian Style

Kita, Akinobu, Hidehiko Okazawa, Katsuya Sugimoto, Nobuyuki Kosaka, Eiji Kidoya, and Tetsuya Tsujikawa. 2023. "Specific Binding Ratio Estimation of [123I]-FP-CIT SPECT Using Frontal Projection Image and Machine Learning" Diagnostics 13, no. 8: 1371. https://doi.org/10.3390/diagnostics13081371

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop