You are currently viewing a new version of our website. To view the old version click .
Diagnostics
  • Article
  • Open Access

26 September 2022

Anterior Cruciate Ligament Tear Detection Based on Deep Convolutional Neural Network

and
Vellore Institute of Technology, Chennai 600127, India
*
Author to whom correspondence should be addressed.
This article belongs to the Section Machine Learning and Artificial Intelligence in Diagnostics

Abstract

Anterior cruciate ligament (ACL) tear is very common in football players, volleyball players, sprinters, runners, etc. It occurs frequently due to extra stretching and sudden movement and causes extreme pain to the patient. Various computer vision-based techniques have been employed for ACL tear detection, but the performance of most of these systems is challenging because of the complex structure of knee ligaments. This paper presents a three-layered compact parallel deep convolutional neural network (CPDCNN) to enhance the feature distinctiveness of the knee MRI images for anterior cruciate ligament (ACL) tear detection in knee MRI images. The performance of the proposed approach is evaluated for the MRNet knee images dataset using accuracy, recall, precision, and the F1 score. The proposed CPDCNN offers an overall accuracy of 96.60%, a recall rate of 0.9668, a precision of 0.9654, and an F1 score of 0.9582, which shows superiority over the existing state-of-the-art methods for knee tear detection.

1. Introduction

Deep learning plays a crucial role in distinct medical image processing applications such as image reconstruction, lesion and tissue segmentation, and the characterization and disease detection of medical abnormality [1,2]. Deep learning is important in disease detection in clinical radiology, as it diminishes the bias and errors that occur due to fatigue and distraction. In recent years, deep learning is chiefly used for the detection of breast mammogram masses [3]; opacities, cardiomegaly, and lung nodules in chest radiographs; and interstitial lung disease and lung nodules in chest CT. However, the application of deep learning is rarely addressed for MRI images because of the complexity of the analysis and the processing of multiple sections of MRI images [4,5].
The anterior cruciate ligaments (ACLs) of the knee connect and stabilize the femur to the tibia bone of the leg. The knee consists of four ligaments: two internal ligaments known as anterior cruciate ligament (ACL) and posterior cruciate ligament (PCL) and two outside ligaments known as the medial collateral ligament and lateral collateral ligament. ACL tear injury is most common in sports persons and occurs due to sudden jerks, accidents, extra stretching, running, etc. [6,7]. The symptoms of ACL tear injury include swelling, pain, knee deformation, and difficulty in walking [3,4]. A radiologist’s job is to use radiological images to discover injuries such as torn ACLs. Manually interpreting knee ACL injuries, meniscus tears, and knee cartilage abnormalities from radiological pictures is a tedious task [8,9]. Different tears may cause knee osteoarthritis, osteoporosis, or knee joint replacement. Physical tests and biomarkers, X-rays, computed tomography (CT), mammography, ultrasound imaging, and magnetic resonance imaging (MRI) are all used to identify an ACL tear in the knee. As the ACL is not visible on a basic X-ray, MRI is the best option for identifying ACL tears [10]. Sprains and partial tears of the ACL can be distinguished from the complete and partial tears of the meniscus via MRI. An ACL is often a narrow band of low signal intensity that runs from the femoral end to the apex and can be visible in a single slice or numerous slices depending on the rotation of the scanning. ACL tears must be read in the coronal, sagittal, and axial planes to obtain the full picture of the ACL tear [11]. ACL injuries can be detected by analyzing the structure and orientation of the ligaments in an MRI image. ACL tears show changes in contour, and discontinuity and signal irregularity within the injured ligament [12]. Deep learning algorithms for image analytics have been widely applied in the medical imaging arena in recent years to handle problems such as segmentation, detection, classification, and diagnosis without the intervention of a radiologist. In the field of categorization and representation learning, the use of CNNs has received considerable attention in recent years. CNNs are effective classifiers that have excellent accuracies in a wide range of applications and have a large number of free parameters. However, many deep-learning frameworks provide a poor feature representation of MRI images because of the improper selection of network hyperparameters.
This paper presents an ACL tear detection method based on a compact parallel deep convolutional neural network (CPDCNN) using knee MRI images. The significant contributions of the paper are summarized as follows:
  • The investigation of CPDCNN for improving the deep feature discrimination to represent the complex features of knee MRI images for ACL tear detection;
  • The performance estimation of the proposed CPDCNN using the public MRNet dataset based on various evaluation metrics such as accuracy, precision, recall rate, and the F1 score.
The rest of the paper is structured as follows: Section 2 includes prior work on ACL tear detection of knee. Section 3 provides brief information on the proposed CPDCNN for ACL tear detection. Section 4 gives a detailed information about experimental results and parameter configurations. Section 5 depicts the discussions and findings from the result. Finally, Section 6 concludes of the proposed work and provides the future scope of the proposed method to improve its performance.

3. Proposed Methodology

Most of the traditional sequential DCNN networks consider the same filter kernel size for convolutional operation in all the layers connected one after another. However, the selection of a specific filter kernel is challenging because smaller filter windows (e.g., 3 × 3 pixels) provide a better representation of finer textures but may neglect larger texture variations in an MRI image, whereas larger filter windows (e.g., 7 × 7 pixels) provide a superior representation of coarse textures but may neglect finer texture variations.
This paper presents a compact parallel deep convolutional neural network (CPDCNN) that consists of three parallel segments of a DCNN with differentially sized filter kernels, as shown in Figure 1. The first parallel layer includes a three-layered DCNN architecture with 3 × 3 pixels convolutional filter kernels at each convolutional layer. The second and third parallel layers consist of 5 × 5 and 7 × 7 pixels convolutional filter kernels, respectively, at each convolutional layer. The proposed CPDCNN is aimed to provide a better fine and coarse texture representation of knee MRI images and minimize the computational complexity of the network. The CNN layer in each parallel segment consists of a convolutional layer (CL), a rectified linear unit layer (ReLU), a maximum pooling layer (MP), a fully connected layer, and a classification layer. Each parallel segment consists of a total of 32, 64, and 128 convolutional filters at the first, second, and third CNN layers, respectively. Figure 2 illustrates the representation of the first CNN layer of the proposed CPDCNN.
Figure 1. Flow diagram of proposed DCNN-based ACL tear detection.
Figure 2. Representation of the first parallel layer of the proposed scheme.

3.1. Convolutional Layer

In the convolutional layer, a leg knee MRI image is convolved with convolutional filters. A convolutional layer gives a better spatial and temporal relationship between the local regions of the leg knee image signal. It provides local connectivity between the various pixels that help to learn the discriminative attributes of changes in the frequency and amplitude of the leg knee image signal indicating a tear injury. In the convolutional layer, the original knee MRI is convolved with the N number of convolutional filters. During the convolutional operation, the filter window of w × w size is multiplied with the w × w size local region of the image using element multiplication.
To capture the local connections of the leg knee image signal texture, the convolutional filter window can be moved over the complete knee MRI image. To prevent the window’s edges from disappearing, one pixel is retained in the window. Throughout the training phase, filter weights are continually updated and initialized at random. The original spectrogram is zero-padded with one pixel all around the edges to keep the image’s original size. N convolutional feature maps are produced through this convolutional procedure. Different textural characteristics of the cloth material may be learned by each filter. The intrinsic textural characteristics of the fabric material are represented by convolutional layers. Over the Whole image, the convolutional kernel advances one pixel at a time. Six convolutional kernels with a size of m × n are chosen for implementation. The initialization of each filter kernel is random. Each value on the map is called a neuron. Equations (1) and (2) provide the convolutional process for leg and knee images I c o n v using filter F. The original image I ( m , n ) having dimensions of R × C is convolved using the filter kernel F with the stride of one pixel over the rows and columns of the knee MRI image.
I c o n v ( m , n ) = I ( m , n ) F
I c o n v ( m , n ) = i = 1 R j = 1 C I ( i , j ) F ( i m , j n )

3.2. ReLU Layer

The convolutional feature map’s nonlinear qualities are enhanced by the rectified linear unit. The convolutional layer’s negative values might cause the features’ nonlinear characteristics to deteriorate. The size of the feature mappings in this layer is identical to those in the convolutional layer. The negative values in convolutional feature maps are rounded to 0 in this layer, and all non-negative values are left alone to create the ReLU feature map ( I R e L U ) as given in Equation (3).
I R e L U ( i , j ) = max i = 1 : R ,   j = 1 : C { 0 , I conv ( i , j ) }

3.3. Batch Normalization

The BN layer normalizes the output of the convolutional layer to reduce the internal co-variance change in the network and minimize the inter-dependency between the various layers.

3.4. Maximum Pooling Layer

By ignoring the less important features, the pooling layer assists in gathering the key characteristics from the ReLU layer feature maps. The maximum of the local window is chosen as the salient characteristic in maximum pooling. Additionally, pooling aids in reducing feature dimensions. Feature maps are scaled by a factor of (1/M) for the largest pooling window, which has M × M dimensions. For the nonoverlapping window of size M × M, pooling is performed. Equation (4) may be used to calculate the maximum pooling (IMP) for the ReLU layer output with a dimension of R × C.
I M P ( i , j ) = max i = 1 : R 1 ,   j = 1 : C 1 I R e L U ( i : i + 1 , j : j + 1 )

3.5. Fully Connected Layer

In a fully connected layer, each neuron of one layer is connected to all other neurons of the other layers to give a deeper representation of the leg knee image.

3.6. Softmax Layer

To determine the class of an unknown sample, a classification layer called Softmax is used. The probabilities for each class are provided, and the class label with the highest probability is chosen as the output class. Since the Adam optimization technique requires less memory and has simple computing requirements, it is utilized to learn the CPDCNN algorithm. It has a learning rate of 0.001, decay rates of 0.9 and 0.999, and a small positive parameter ε = 10 8 to avoid division by zero [19].

4. Experimental Results

4.1. System Configurations and Dataset

The proposed system was implemented using Python-OpenCV programming on a personal computer having a core i3 processor with a speed of 2.64 GHz, 4 GB RAM, and a Windows environment. The performance of the proposed CPDCNN was evaluated on the MRNet knee joint MRI dataset [20]. The MRNet dataset consists of MRI images in sagittal T2, coronal T1, and axial PD views with multiple frames that consist of ACL and meniscal tears. We selected specific frames from the dataset in the sagittal T2 view, which consists of a complete view of normal and tear ligaments. Each image in the dataset has a dimension of 256 × 256 pixels. The sample images from the dataset are illustrated in Figure 3. We selected 845 normal and 450 abnormal samples to form the dataset. Out of the total database, 70% of the data were used for training, and 30% of the data were used for testing.
Figure 3. Sample images from the MRNet database.

4.2. Performance Metrics

The outcomes of the proposed scheme were validated using various quantitative and qualitative metrics such as precision, recall, accuracy, and the F1 score. Equations (5)–(8) are used to calculate precision, recall, accuracy, and the F1 score, respectively.
P r e c i s i o n = T P T P + F P
R e c a l l = T N T N + F N
A c c u r a c y   ( % ) = T P + T N T P + T N + F P + F N × 100
F 1 S c o r e = 2 P r e c i s i o n R e c a l l P r e c i s i o n + R e c a l l

4.3. Network Parameters

The parameter specifications and feature maps of the different layers of the proposed CPDCNN knee tear detection system are given in Table 1. The proposed CPDCNN provided 393,216 trainable parameters per parallel arm and a total of 786,434 trainable parameters for knee tear detection.
Table 1. Parameter specification of proposed CPDCNN.
Table 2 provides the experimental results for the proposed CPDCNN for various CNN layers at each parallel arm. The use of a three-layered CPDCNN provided a better representation of the fine and coarse textures of the knee MRI image, thus helping to characterize such tear injuries from normal knee MRI images. The proposed CDCNN provided accuracy values of 91.34%, 94.48%, and 96.60% for CPDCNNs that included, respectively, one CNN with 32 filters, two CNNs with 32 and 64 filters, and three CNNs with 32, 64, and 128 filters for convolutional operation. Increasing the CNN layers in the parallel arm showed significant improvement in tear detection accuracy because of improvement in deep feature representation. Figure 4 and Figure 5 illustrate the performance of the proposed CPDCNN for different CNNs at each parallel arm.
Table 2. Performance of the proposed CPDCNN for different CNN layers.
Figure 4. Accuracy for proposed CPDCNN for various layers.
Figure 5. Recall, precision, and F1 score for proposed CPDCNN for various layers.
The outcomes of the proposed CPDCNN-based ACL tear detection were evaluated for the different learning algorithms such as Adam, stochastic gradient descent momentum (SGDM), and root-mean-square propagation (RMSProps) algorithm, as described in Table 3. It had 96.60% accuracy for the 3 × 3 pixels convolutional kernel, which is superior to other windows. The 3 × 3 pixels window provides better spatial connectivity of the local regions of the MRI image that help to characterize the normal and tear textures of the knee MRI image. It provided 96.60%, 95.88%, 94.48%, and 92.92% accuracy values for the 3 × 3 pixels, 5 × 5 pixels, 7 × 7 pixels, and 9 × 9 pixels convolutional filter kernels, respectively. The CPDCNN-Adam provided 96.60% accuracy, which is superior to CPDCNN-SGDM (95.88) and CPDCNN-RMSProps (94.48%). The F1 score represents a good balance between the quantitative (recall) and qualitative (precision) performance of the CPDCNN for knee tear detection. The F1 scores of the proposed CPDCNN were 0.9610, 0.9589, and 0.9415 for Adam, SGDM, and RMSProps learning strategies, respectively.
Table 3. Performance comparison of proposed system with previous ACL tear detection approaches.

5. Discussion

The performance of the proposed system was compared with previous deep-learning-based ACL tear detection techniques, as shown in Table 3. The proposed system showed significant improvement in ACL tear detection, compared with previous deep-learning-based approaches for ACL tear detection. The proposed CPDCNNN showed an improvement of 1.68%, 3.42%, 6.09%, 7.15%, and 18.86% in ACL tear detection accuracy over DCNN [13], ResNet18 [22], Efficient Net [17], MLP [14], and ResNet50 [21], respectively, for ACL tear detection. The proposed CPDCNN provided a total of 786434 trainable parameters, which are fewer compared with ResNet50 [21], ResNet18 [22], and ResNet [23]. The proposed network’s fewer trainable parameters provide faster training, testing, and implementation feasibility on standalone devices based on microprocessors or microcontrollers.

6. Conclusions and Future Scope

Thus, in the proposed system, CPDCNN was presented to improve feature distinctiveness, thus improving the representation capability of the complex knee MRI texture for ACL tear detection using knee MRI images. CPDCNN helps to discriminate between a torn texture and normal textures of the ACL region in knee MRI images and provides better local and global feature representation capability. The proposed CPDCNN had 96.60% accuracy on the Knee MRI dataset, and it helps to minimize the complexity of deep learning frameworks and the problem of hyperparameter tuning. The results of this study revealed that the proposed system had better performance than the previous state-of-the-art deep-learning-based methods for ACL tear detection. In the future, the performance of the proposed approach can be improved by using an augmented dataset. The CPDCNN provided a total of 786,434 trainable parameters, which are lower than the existing state-of-the-art methods, thus helping to minimize the computational complexity of the network. The proposed system yielded 96.60% accuracy for the three-layered CPDCNN for ACL tear detection. This method can be extended to develop a generalized framework for the detection of multiple types and multi-stage knee tear detection using datasets with higher volumes of data.

Author Contributions

K.S.: Conceptualization, methodology, resources, writing—review and editing, supervision; K.J.: Methodology, Validation, writing—review and editing. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The dataset used for the work can be found at https://stanfordmlgroup.github.io/competitions/mrnet/ accessed on 1 January 2021.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Liu, F.; Jang, H.; Kijowski, R.; Bradshaw, T.; McMillan, A.B. Deep learning MR imaging–based attenuation correction for PET/MR imaging. Radiology 2018, 286, 676–684. [Google Scholar] [CrossRef] [PubMed]
  2. Zhu, B.; Liu, J.Z.; Cauley, S.F.; Rosen, B.R.; Rosen, M.S. Image reconstruction by domain-transform manifold learning. Nature 2018, 555, 487–492. [Google Scholar] [CrossRef] [PubMed]
  3. Berk, N.; Pedoia, V.; Majumdar, S. Use of 2D U-Net convolutional neural networks for automated cartilage and meniscus segmentation of knee MR imaging data to determine relaxometry and morphometry. Radiology 2018, 288, 177–185. [Google Scholar]
  4. Liu, F.; Zhou, Z.; Jang, H.; Samsonov, A.; Zhao, G.; Kijowski, R. Deep convolutional neural network and 3D deformable approach for tissue segmentation in musculoskeletal magnetic resonance imaging. Magn. Reson. Med. 2018, 79, 2379–2391. [Google Scholar] [CrossRef] [PubMed]
  5. Zhou, Z.; Zhao, G.; Kijowski, R.; Liu, F. Deep convolutional neural network for segmentation of knee joint anatomy. Magn. Reson. Med. 2018, 80, 2759–2770. [Google Scholar] [CrossRef]
  6. Liu, F.; Zhou, Z.; Samsonov, A.; Blankenbaker, D.; Larison, W.; Kanarek, A.; Lian, K.; Kambhampati, S.; Kijowski, R. Deep learning approach for evaluating knee MR images: Achieving high diagnostic performance for cartilage lesion detection. Radiology 2018, 289, 160–169. [Google Scholar] [CrossRef]
  7. Joseph, F.D.; Ledsam, J.R.; Romera-Paredes, B.; Nikolov, S.; Tomasev, N.; Blackwell, S.; Askham, H.; Glorot, X.; O’Donoghue, B.; Visentin, D.; et al. Clinically applicable deep learning for diagnosis and referral in retinal disease. Nat. Med. 2018, 24, 1342–1350. [Google Scholar]
  8. John, A.; González, F.A.; Ramos-Pollán, R.; Oliveira, J.L.; Lopez, M.A.G. Representation learning for mammography mass lesion classification with convolutional neural networks. Comput. Methods Programs Biomed. 2016, 127, 248–257. [Google Scholar]
  9. Mark, C.; Bilbily, A.; Colak, E.; Dowdell, T.; Gray, B.; Perampaladas, K.; Barfett, J. Training and validating a deep convolutional neural network for computer-aided detection and classification of abnormalities on frontal chest radiographs. Investig. Radiol. 2017, 52, 281–287. [Google Scholar]
  10. Marios, A.; Christodoulidis, S.; Ebner, L.; Christe, A.; Mougiakakou, A. Lung pattern classification for interstitial lung diseases using a deep convolutional neural network. IEEE Trans. Med. Imaging 2016, 35, 1207–1216. [Google Scholar]
  11. Lee, J.K.; Yao, L.; Phelps, C.T.; Wirth, C.R.; Czajka, J.O.H.N.; Lozman, J. Anterior cruciate ligament tears: MR imaging compared with arthroscopy and clinical tests. Radiology 1988, 166, 861–864. [Google Scholar] [CrossRef] [PubMed]
  12. Partezani, H.P.V.; Helito, C.P.; Rodrigues, M.B. Anterolateral ligament MRI of the knee in ACL injuries: MRI abnormalities association with instability. Eur. Radiol. 2022, 2022, 1–9. [Google Scholar]
  13. Nicholas, B.; Rajpurkar, P.; Ball, R.L.; Irvin, J.; Park, A.; Jones, E.; Bereket, M.; Patel, B.N.; Yeom, K.W.; Shpanskaya, K.; et al. Deep-learning-assisted diagnosis for knee magnetic resonance imaging: Development and retrospective validation of MRNet. PLoS Med. 2018, 15, e1002699. [Google Scholar]
  14. Lai, Z.; Deng, H. Medical Image Classification Based on Deep Features Extracted by Deep Model and Statistic Feature Fusion with Multilayer Perceptron. Comput. Intell. Neurosci. 2018, 2018, 2061516. [Google Scholar] [CrossRef]
  15. Liu, F.; Guan, B.; Zhou, Z.; Samsonov, A.; Rosas, H.; Lian, K.; Sharma, R.; Kanarek, A.; Kim, J.; Guermazi, A.; et al. Fully automated diagnosis of anterior cruciate ligament tears on knee MR images by using deep learning. Radiol. Artif. Intell. 2019, 1, 180091. [Google Scholar] [CrossRef] [PubMed]
  16. Emre, Ö.; Akdoğan, V.; Korkmaz, M.; Orhan, E.R. Automatic Segmentation of Meniscus in Multispectral MRI Using Regions with Convolutional Neural Network (R-CNN). J. Digit. Imaging 2020, 33, 916–929. [Google Scholar]
  17. Tsai, C.; Kiryati, N.; Konen, E.; Eshed, I.; Mayer, A. Knee Injury Detection using MRI with Efficiently-Layered Network (ELNet). arXiv 2020, arXiv:2005.02706. [Google Scholar]
  18. Christoph, G.; Marbach, G.; Civardi, F.; Fucentese, S.F.; Fritz, J.; Sutter, R.; Pfirrmann, C.W.A.; Fritz, B. Deep Convolutional Neural Network–Based Diagnosis of Anterior Cruciate Ligament Tears: Performance Comparison of Homogenous Versus Heterogeneous Knee MRI Cohorts with Different Pulse Sequence Protocols and 1.5-T and 3-T Magnetic Field Strengths. Investig. Radiol. 2020, 55, 499. [Google Scholar]
  19. Kingma, D.P.; Adam, J.B. A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  20. MRNet Dataset. Available online: https://stanfordmlgroup.github.io/competitions/mrnet/ (accessed on 5 August 2021).
  21. Kara, A.C.; Hardalaç, F. Detection and classification of knee injuries from MR images using the MRNet dataset with progressively operating deep learning methods. Mach. Learn. Knowl. Extr. 2021, 3, 1009–1029. [Google Scholar] [CrossRef]
  22. David, A.; McGuinness, K.; Smeaton, A.F. A comparative study of existing and new deep learning methods for detecting knee injuries using the MRNet dataset. In 2020 International Conference on Intelligent Data Science Technologies and Applications (IDSTA); IEEE: Piscataway, NJ, USA, 2020; pp. 149–155. [Google Scholar]
  23. Ismail, I.; Anwar, S.M.; Torigian, D.A.; Bagci, U. Deep learning for musculoskeletal image analysis. In 2019 53rd Asilomar Conference on Signals, Systems, and Computers; IEEE: Piscataway, NJ, USA, 2019; pp. 1481–1485. [Google Scholar]
  24. Awan, M.J.; Rahim, M.S.M.; Salim, N.; Mohammed, M.A.; Garcia-Zapirain, B.; Abdulkareem, K.H. Efficient detection of knee anterior cruciate ligament from magnetic resonance imaging using deep learning approach. Diagnostics 2021, 11, 105. [Google Scholar] [CrossRef] [PubMed]
  25. Alexia, T.; Lassalle, L.; Zille, P.; Guillin, R.; Pluot, E.; Adam, C.; Charachon, M.; Brat, H.; Wallaert, M.; d’Assignies, G.; et al. Deep learning to detect anterior cruciate ligament tear on knee MRI: Multi-continental external validation. Eur. Radiol. 2022, 2022, 1–10. [Google Scholar]
  26. Yusuke, M.; Akagi, R.; Maki, S.; Shiko, Y.; Tozawa, R.; Kimura, S.; Yamaguchi, S.; Kawasaki, Y.; Ohtori, S.; Sasho, T. Automated detection of anterior cruciate ligament tears using a deep convolutional neural network. BMC Musculoskelet. Disord. 2022, 23, 1–7. [Google Scholar]
  27. Vansh, K.; Tyagi, N.; Manocha, B.; Arora, A.; Roy, S.; Nagrath, P. Detection of anterior cruciate ligament tear using deep learning and machine learning techniques. In Data Analytics and Management; Springer: Singapore, 2021; pp. 9–22. [Google Scholar]
  28. Rizk, B.; Lassalle, L.; Pluot, E.; Guillin, R.; Zille, P.; Tran, A.; D’Assignies, G. Performance Comparison of a Deep Learning Algorithm and Human Readers on Detection of Anterior Cruciate Ligament Tear. In Seminars in Musculoskeletal Radiology; Thieme Medical Publishers, Inc.: New York, NY, USA, 2022; Volume 26, p. A121. [Google Scholar]
  29. Shaju, J.K.; Joshy, N.A.; Singh, A.R.; Jadhav, R. SmartACL: Anterior Cruciate Ligament Tear Detection by Analyzing MRI Scans. In Innovations in Computational Intelligence and Computer Vision; Springer: Singapore, 2022; pp. 373–381. [Google Scholar]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.