Next Article in Journal
Mindless Memorization Booster: A Method to Influence Memorization Power Using Attention Induction Phenomena Caused by Visual Interface Modulation and Its Application to Memorization Support for English Vocabulary Learning
Next Article in Special Issue
RED-CNN: The Multi-Classification Network for Pulmonary Diseases
Previous Article in Journal
Two Sound Field Control Methods Based on Particle Velocity
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Enhanced Automatic Morphometry of Nerve Histological Sections Using Ensemble Learning

1
Department of Biomedical Engineering, Jordan University of Science and Technology, Irbid 22110, Jordan
2
Neural Engineering Center, Case Western Reserve University, Cleveland, OH 44106, USA
*
Author to whom correspondence should be addressed.
Electronics 2022, 11(14), 2277; https://doi.org/10.3390/electronics11142277
Submission received: 19 June 2022 / Revised: 13 July 2022 / Accepted: 16 July 2022 / Published: 21 July 2022

Abstract

:
There is a need for an automated morphometry algorithm to facilitate the otherwise labor-intensive task of the quantitative histological analysis of neural microscopic images. A benchmark morphometry algorithm is the convolutional neural network Axondeepseg (ADS), which yields a high segmentation accuracy for scanning and transmission electron microscopy images. Nevertheless, it shows decreased accuracy when applied to optical microscopy images, and it has been observed to yield sizable false positives when identifying small-sized neurons within the slides. In this study, ensemble learning is used to enhance the performance of ADS by combining it with the paired image-to-image translation algorithm PairedImageTranslation (PIT). Here, 120 optical microscopy images of peripheral nerves were used to train and test the ensemble learning model and the two base models individually for comparison. The results showed weighted pixel-wise accuracy for the ensemble model of 95.5%, whereas the ADS and PIT yielded accuracies of 93.4% and 90%, respectively. The automated measurements of the axon diameters and myelin thicknesses from the manually marked ground truth images were not statistically different (p = 0.05) from the measurements taken from the same images when segmented using the developed ensemble model, while they were different when they were measured from the segmented images by the two base models individually. The automated measurement of the G ratios indicated a higher similarity to the ground truth testing images for the ensemble model in comparison with the individual base models. The proposed model yielded automated segmentation of the nerve slides, which were sufficiently equivalent to the manual annotations and could be employed for axon diameters and myelin thickness measurements for fully automated histological analysis of the neural images.

1. Introduction

Communication between the central nervous system (CNS) and the body is established through the peripheral nervous system (PNS). Any problem that interrupts this communication can lead to various motor and sensory deficits (e.g., multiple sclerosis and neurodegenerative diseases). Such deficits can be attributed to several factors, such as demyelination, axonal loss, or a mixture of both. Motor and sensory disabilities are correlated with the severity of these defects [1,2]. Therefore, the ability to efficiently image, identify, and quantify nerve histological structures would allow researchers to track the progression and severity of these defects, and to assess the impact of therapeutic drugs [3,4].
In an attempt to automate this otherwise labor-intensive analysis, multiple traditional image segmentation algorithms have been applied for myelin and axon segmentation and labeling. Examples of these algorithms include thresholding, region growing, morphological operations, watershed, and active contours [5,6,7,8,9,10,11]. However, such methods did not gain much popularity due to their limited scope of application and accuracy [12].
In the last few years, deep learning methods, particularly convolutional neural networks (CNNs), have shown great capabilities in medical image segmentation [13]. A recent study presented the open-source software Axondeepseg (ADS). This software implements the CNN approach to segment the nerve axon and myelin with a pixel-wise accuracy that reached up to 95% [14]. The performance of this software has been reported for images collected from SEM and TEM, but not for the cheaper, widely used optical microscopy, raising the need for further exploration. Additionally, optical microscopy is prone to a high level of variability due to the preparation and staining protocol used, which may inherently lead to different variations that are not usually present in SEM and TEM images.
Another deep learning method for image transformation applications is paired image-to-image translation [15]. It is commonly used in computer vision where an image from one domain is converted into a corresponding image in another domain, given sufficient training data [16]. It is accustomed to learning a joint distribution of images in different domains by using images from the marginal distributions in individual domains [17]. This capability could be exploited in neural segmentation tasks, with the two image domains being the original images before the markings and the same image after the segmentation or marking. This approach has not been applied in the neural image segmentation task before.
Combining the performance of the two aforementioned techniques is possible with the use of ensemble learning, where multiple learning algorithms are combined to obtain a better predictive performance than what could be obtained from any of the constituent learning algorithms alone [18]. In this study, the performance of the ADS and a PIT-based classification software (PairedImageTranslation [15]) are investigated individually for segmenting optical microscopy images of nerve slices, then ensemble learning is employed to enhance the segmentation performance by combining the ADS and PIT models. The resulting segmented images are then exploited for the automated measurement of morphological parameters, such as the axon diameter and myeline thickness, and are contrasted with the measurements that were based on manually annotated ground truth images.

2. Materials and Methods

2.1. Datasets

The images used in this study were obtained from sciatic nerves harvested from four mongrel dogs as a part of a neural interface project. Following the perfusion process, the nerve was cut in sections at 0.5–1.0 μm. Toluidine Blue staining (1% Toluidine Blue and 2% Borate) was used [19] to stain the myelin sheath of the nerve axons.
The prepared slides were then imaged using optical microscopy with a resolution of 0.15 µm/pixel. All of the original animal experiments that produced the used images were approved by the Institutional Animal Care and Use Committee (IACUC) of the Case Western Reserve University (Cleveland, OH, USA). The data set consisted of 120 images of a size of 512 × 512 pixels. These images were distributed into 72 images for training, 24 for validation, and 24 for testing.

2.2. Ground Truth Labeling

Ground truth labels were created in two steps, as follows. (i) Inner and outer perimeters of the myelin were labeled manually by initially adjusting the image contrast, performing an automated edge-detection algorithm, and then visually inspecting the image to retain the detected borders while erasing noise lines. (ii) Images were then edited using the Open Source image editor GIMP to correct remaining incomplete boundaries and to fill the axon, myelin, and background regions with black, gray, and white colors, respectively. Figure 1 shows examples of the original images and their corresponding labeled images.

2.3. Network Architecture

Both the ADS and PIT frameworks originate from the U-net model [20]. This model is structured in a U shape, which contains two paths: the contracting path and the expanding path. In the contractile path, images are encoded into feature maps at several CNN layers. At each CNN layer, the number of feature maps is doubled, whereas the dimension of the image is reduced by half. The expanding path of the U-Net model works in the exact opposite way as that of the contractile path. At each CNN layer, the number of feature maps is halved, whereas the dimension of the image is doubled until it reaches the dimension of the original image [12,15].

2.3.1. ADS Network Architecture

Each contracting and expanding path of ADS consists of three blocks, and each block has three convolutional layers. In the first block, the convolutional layers use 5 × 5 kernels, while the convolutional layers in the remaining two blocks use 3 × 3 kernels. Each block in the contracting path ends with a convolution of stride 2 for dimensionality reduction. Each strided convolution layer has a corresponding up-convolution layer in the expansion path. The number of channels in the first block starts with 16 and doubles after each block, and then decreases by half in each block of the expansion path. The activation function used in all of the convolutional layers is rectified linear units (ReLU).

2.3.2. PIT Network Architecture

Each contracting and expanding path of PIT consists of four blocks, and each block has two convolutional layers. All of the blocks use 3 × 3 kernels. Each block in the contracting path ends with a convolution of stride 2 for dimensionality reduction. Each strided convolution layer has a corresponding up-convolution layer in the expansion path. The number of channels in the first block starts with 32 and doubles after each block, and then decreases by half in each block of the expansion path. The activation function used in all of the convolutional layers is rectified linear units (ReLU). The output of each contracting path is connected directly to the corresponding expanding path. This is called a skip connection. The idea of skip connections is to transfer the raw information directly to the final output. This helps mitigate the vanishing-gradient problem and accelerates learning.

2.4. Ensemble Learning

Ensemble learning, also called committee-based learning and a multi-classifier system, uses two or more base models to enhance the classification performance and improve the robustness and quality of the individual models. First, a set of individual models are trained separately. Then, the classification results of the individual models are aggregated by one of the widely used ensemble strategies, such as voting, bagging, boosting, stacking, Bayes optimal classifier, and winner-take-all [21]. In the presented study, we aggregated the results via averaging the probability matrix as described in the following section, which we found to be sufficient to yield the anticipated improvement, as the two base models are distinct in their approach.

Ensemble Learning Pipeline

In the learning step, the training and validation images were fed into both the ADS and PIT models, and then both trained models were evaluated on the testing of four intact images and the two probability matrices that resulted after the Softmax function were averaged to obtain a single probability matrix. Finally, argmax function was used to obtain the prediction matrix, which was then compared with the testing ground truth (manually labeled images) using evaluation metrics. A flowchart of the entire learning and testing process is provided in Figure 2.

2.5. Morphological Analysis

The main morphological parameters of interest are the G-ratio, axonal diameter, and myelin thickness [22]. The G-ratio is used to assess axonal myelination and reflects the efficiency of neural pathways to conduct neural signals [23]. It is estimated by dividing the axonal diameter by the myelinated fiber diameter. The axon shapes were assumed to be circular; therefore, the axon and myelinated axon (axon + mylein) diameters were calculated by equating the measured areas of the original myelinated axon to the area of two concentric circles, as depicted in Figure 3.

2.6. Evaluation Metrics

To evaluate the performance of the trained model for the segmentation task, several evaluation metrics were used. Dice coefficient and pixel-wise accuracy for axon, myelin, and background were used to indicate their segmentation quality. The dice coefficient was calculated between the segmented image and the ground truth image for the testing image set. It is defined as follows:
Dice = 2 A B A + B
where A is the number of pixels of a given class that are present in the segmented image, B is the number of pixels of a given class in the ground truth image, and (A ∩ B) is the number of pixels of the same type that are present in both images (correctly segmented pixels).
Pixel accuracy is a semantic segmentation metric that denotes the percent of pixels that are accurately classified in the image. It is defined as follows:
Accuracy = Number   of   adequately   classified   pixels Total   number   of   pixels   in   the   image
As the number of pixels per class is imbalanced, the total accuracy performance classes are calculated as a weighted pixel-wise accuracy, yielding weights of 1, 2.58, and 3.69 for the background, myelin, and axon, respectively.
Precision and sensitivity (also known as recall) for the both axon and myelin were calculated to assess the ability to detect true axon, myelin, and background regions, while avoiding false positives [12]. They were calculated for each pixel type as follows:
Precision = T P A T P A + F P A
Sensitivity = T P A T P A + F N A  
where T P A represents the count of pixels correctly classified as pixel A, F P A represents the count of the other pixel types incorrectly classified as pixel type A, and F N A represents the count of the pixels of type A that are incorrectly classified as pixels of other types. These two metrics were calculated for the three pixel types (i.e., “type A” indicates either axon, myeline, or background).

3. Results

3.1. Quantitative and Qualitative Analysis of Pixel-Wise Segmentation

The performance of the proposed model was quantitatively assessed by classifying each pixel of the analyzed image into three possible categories, namely axon, myeline, or background. Table 1 lists the evaluation metrics computed on the outputs of the ADS, PIT, and ensemble models. The weighted pixel-wise accuracy of the ensemble model was 95.5%, whereas the ADS and PIT models achieved a weighted pixel-wise accuracy of 93.4% and 90%, respectively.
The axon dice of the ensemble model reached 94.3%, which was higher than that of the ADS by 2%.
The confusion matrices for the three models are shown in Figure 4. The ensemble model had an axon recall of 0.94, which outperformed the two base models by 5% in ADS and 3% in PIT. Myelin recall was also slightly enhanced in the ensemble model at 0.95, compared with that of the ADS model at 0.94.
To express the performance of the developed model qualitatively, Figure 5 shows the segmentation results of three sample images against the original images and their ground truth labeled images. The quality of the segmentation results varied among the three models, which reflects the aforementioned variation in the accuracy values. Overall, ADS shows a superior performance over PIT in classifying the transition borders between different cell types, hence the PIT-classified images appear to be pixilated. Nevertheless, PIT appeared to yield a better outcome when eliminating the artifacts (marked with red ellipses in the rightmost column of Figure 5) and yielded a more uniform marking of the axonal regions that are surrounded by myelin sheaths (marked with red rectangles in Figure 5). The ensembled learning model appeared to combine the individual proficiencies of each of the two base models and yielded a superior overall quality of the segmented images.

3.2. Quantitative Analysis of Automated Morphometry

Classifying the neural images into the three pixel classes does not have a direct utility per se in the histological assessment of the nerve tissue for which the axonal morphometry data are needed. These data encompass the distributions of the axonal diameters and the corresponding myelin thicknesses for the axonal population, along with the relative myelination index known as the G-ratio. Nevertheless, the automated measurements of the diameters and myelin thicknesses are impacted by the segmentation quality, hence a high accuracy segmentation is needed. To assess the utility of the three segmentation models in obtaining these measurements, automated morphometry was applied to the segmented images by the three models, and the results are compared with the morphometric data resulting from the ground truth images. The distributions yielded of the axonal diameter, myeline thicknesses, and G-ratios are shown in Figure 6.
Statistical analysis was performed for any significant difference in the morphological parameters calculated from the ground truth, ensemble, ADS, and PIT models. The analysis was performed on a total of 628 axons contained in the testing images. The normality assumption was checked using Kolmogorov–Smirnov test as the number of axons was >300. The G-ratio values were found to be normally distributed. Thus, a one way ANOVA test was performed followed by a Student’s t-test. The axon diameter and myelin thickness data were found to be not normally distributed (p < 0.05). Thus, the Friedman test was carried out, followed by the two-way Wilcoxon signed-rank test. The statistical significance used in this study was defined as p < 0.05.
Figure 6A shows the G-ratio distributions. All four distributions were found to be normally distributed. Both the ensemble and ADS models yielded G-ratio distributions that were statistically different from the ground truth G-ratios. Nevertheless, the yielded p-Value was higher for the ensemble G-ratios (p = 0.0002), which indicates slightly enhanced G-ratios in the ensemble model. Figure 6B shows the axon diameter distributions. All four distributions were found to be not normally distributed. Statistical analysis of the diameter distributions indicated no significant difference between the ensemble model and ground truth, while being statistically different from the distributions yielded from the base models individually. Figure 6C shows the myelin thickness distributions. All four distributions were found to be not normally distributed, and further analysis indicated no statistical difference between the results from the ensemble model’s segmented images and ground truth images, while the base models yielded statistically different distributions.
Table 2 lists the G-ratios calculated from the ground truth testing images and for the corresponding predicted images of the ADS, PIT, and ensemble models. The average G-ratio for four testing whole slides (1360 × 1024 pixels each) was 0.531 ± 0.029 when calculated from ground truth images. The G-ratio calculated from the ensemble model was the closest to that of the ground truth images at 0.524 ± 0.019, with a percentage difference of 1.4%. However, the average G-ratio calculated from the ensemble model was statistically different from the ground truth results (p = 0.0002).

4. Discussion

The presented study evaluated the performance of an ensemble learning model that combines ADS and PIT in segmenting optical microscopy peripheral nerve images for automated morphometry applications. The weighted pixel-wise accuracy of the proposed model was 95.5%. The maximum pixel-wise accuracy found in the literature for axon and myelin segmentation on optical microscopy images was 82.3% using a deep convolutional encoder-decoder [23] for 32 × 32 pixels images of a mouse spinal cord.
A fully convolutional neural network was previously used for axon diameter distribution and resulted in a high correlation with the ground truth, but its utility was demonstrated on grayscale images of central neural structures using electron microscopy images [24].
Ensemble learning has been explored for in vetro axon segmentation, where U-net with a ResNet-50 encoder architecture was used [17]. Nevertheless, it was applied on axonal phase-contrast images to allow for 3D axonal tracing and not the analysis of high-resolution neural cross-sectional images.
The U-net architecture was used alone for the segmentation task while incorporating a novel dropout, where the Monte-Carlo dropout approach was incorporated to assess the model uncertainty and to select the samples to be annotated for the next iteration [25]. It was able to improve dice values by 10% to up to 90%, with a relatively small training im-age count (<100). In our model, each base model was utilized to produce weak predictive results, then ensemble learning fused the favorable prediction via voting schemes in an adaptive way. Hence the incremental improvement on one of the base models is not guaranteed to lead to the corresponding overall improvement of the ensemble model.
Approaches that do not utilize machine learning were also explored, with accuracy values of single-axon detection up to 82%, and classification of the fiber myelination accuracy of 89% [26]. The segmentation process relied solely on the image processing steps and elliptical shape reconstruction for the input images under different levels of contrast and color saturation. Even though the approach is simple and computationally not expensive, it requires manual adjustment of the color levels by experienced operators, which hinders the full automation of the entire process.
The evaluation metrics calculated using the ensemble model showed improved performance over the two base models separately, except for axon precision (Table 1). The achieved accuracy of the developed ensemble model also surpassed the previously reported accuracy of the ADS base model when segmenting images collected from SEM and TEM [9]. This observation indicates the superiority of the ensemble model to detect true positive axons, which can be attributed to the improved axon regions detected in the PIT model, especially for small axons (Figure 5); hence, negating the need for manual intervention, as it was often required [7,8,11,27].
The axon precision yielded from the ensemble model was lower than that obtained with the ADS model by 1%. The reason for the decreased axon precision in the ensemble model comes from the low axon precision in the PIT model (84.2%) as the PIT predicts segments of the outer borders as axons instead of being predicted as myelin or background. The ensemble learning approach has been shown to address most of these predictions and reflected the superior performance of the ADS at the outer borders.
The utility of the developed ensemble learning model in the automated axonal morphometric analysis was demonstrated with a high consistency between the axonal and myelin thickness measurement based on the ensembled model segmentation, and the measurements that are based on the ground truth testing images. The G-ratios calculated from the ensembled model yielded a percentage difference of 1.4%, which was statistically different from the ground truth. The reason for this observed difference in the G-ratio and not for the axon diameters and myelin thicknesses can be attributed to the fact that the G-ratio calculation incorporates both the axon diameter and the myelin thickness, thereby it can be influenced by two sources of variability cumulatively, resulting in this observed difference from the ground truth images. Nevertheless, the p-Value of the three models’ comparisons indicated less degree of differences between the ensemble model and ground truth than with the other base models individually.

5. Conclusions

In this study, the accuracy of the CNN algorithm ADS in segmenting neural images was improved by combining it with a paired-image-to-image translation algorithm using ensembled learning. The ensemble technique has benefited from the superior specificity of PIT in removing artifacts and in the uniform regional classification within axons. As such, the ensembled learning model yielded improved pixel-wise segmentation accuracy. The utility of the ensemble technique in automating the morphometry analysis of the neural images was demonstrated by measuring the axonal diameters, myelin thicknesses, and G-ratios. The ensemble model yielded statistically similar morphometric parameters (i.e., axon diameter and myelin thickness) to the parameters calculated based on the ground truth images. The G-ratio measurements were statistically different from the ground truth measurements, but were improved over the base models. In future work, universal automated segmentation software can be developed by increasing the number of models used in ensemble learning and training the algorithm on larger more variant datasets that are collected from various species using various preparation and staining protocols.

Author Contributions

Conceptualization, Y.D., M.A.-Z., and D.D.; methodology, Y.D., and M.A.-Z.; software, M.A.-Z.; validation, Y.D. and M.A.-Z.; formal analysis, Y.D., and M.A.-Z.; investigation, Y.D., and M.A.-Z.; resources, Y.D. and M.A.-Z.; data curation, Y.D.; writing—original draft preparation, M.A.-Z.; writing—review and editing, Y.D. and D.D.; visualization, M.A.-Z.; supervision, Y.D., and D.D.; project administration, Y.D. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Deanship of Research at Jordan University of Science and Technology, Irbid, Jordan (Project No. 20200601; Funder ID: 10.13039/501100019004).

Institutional Review Board Statement

All of the original animal experiments that produced the used images were approved by the Institutional Animal Care and Use Committee (IACUC) of the Case Western Reserve University, Cleveland, OH, USA. (Animal protocol numbers 2011-0145 and 2014-0014).

Informed Consent Statement

Not applicable.

Data Availability Statement

All of the images used in this study are available upon request.

Acknowledgments

The authors would like to thank the original neural interface project team members; Thomas E, Eggers, Luis E, Gonzalez-Reyes, Joseph Drain, and Grant A, McCallum for all their contributions to the project that produced the used neural images. Also, we would like to thank the Center of Excellence for Innovative Projects at Jordan University of Science and Technology for their technical support.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Krajewski, K.M.; Lewis, R.A.; Fuerst, D.R.; Turansky, C.; Hinderer, S.R.; Garbern, J.; Kamholz, J.; Shy, M.E. Neurological dysfunction and axonal degeneration in Charcot–Marie–Tooth disease type 1A. Brain 2000, 123, 1516–1527. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Trapp, B.D.; Peterson, J.; Ransohoff, R.M.; Rudick, R.; Mörk, S.; Bö, L. Axonal transection in the lesions of multiple sclerosis. N. Engl. J. Med. 1998, 338, 278–285. [Google Scholar] [CrossRef] [PubMed]
  3. Sachs, H.H.; Bercury, K.K.; Popescu, D.C.; Narayanan, S.P.; Macklin, W.B. A new model of cuprizone-mediated demyelination/remyelination. ASN Neuro 2014, 6, 1759091414551955. [Google Scholar] [CrossRef] [PubMed]
  4. Pfeifenbring, S.; Nessler, S.; Wegner, C.; Stadelmann, C.; Brück, W. Remyelination after cuprizone-induced demyelination is accelerated in juvenile mice. J. Neuropathol. Exp. Neurol. 2015, 74, 756–766. [Google Scholar] [PubMed] [Green Version]
  5. Romero, E.; Cuisenaire, O.; Denef, J.-F.; Delbeke, J.; Macq, B.; Veraart, C. Automatic morphometry of nerve histological sections. J. Neurosci. Methods 2000, 97, 111–122. [Google Scholar] [PubMed] [Green Version]
  6. Zhao, X.; Pan, Z.; Wu, J.; Zhou, G.; Zeng, Y. Automatic identification and morphometry of optic nerve fibers in electron microscopy images. Comput. Med. Imaging Graph. 2010, 34, 179–184. [Google Scholar] [PubMed]
  7. Bégin, S.; Dupont-Therrien, O.; Bélanger, E.; Daradich, A.; Laffray, S.; de Koninck, Y.; Côté, D.C. Automated method for the segmentation and morphometry of nerve fibers in large-scale CARS images of spinal cord tissue. Biomed. Opt. Express 2014, 5, 4145–4161. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  8. More, H.L.; Chen, J.; Gibson, E.; Donelan, J.M.; Beg, M.F. A semi-automated method for identifying and measuring myelinated nerve fibers in scanning electron microscope images. J. Neurosci. Methods 2011, 201, 149–158. [Google Scholar] [CrossRef] [PubMed]
  9. Liu, T.; Jurrus, E.; Seyedhosseini, M.; Ellisman, M.; Tasdizen, T. Watershed merge tree classification for electron microscopy image segmentation. In Proceedings of the 21st International Conference on Pattern Recognition (ICPR2012), Tsukuba, Japan, 11–15 November 2012; pp. 133–137. [Google Scholar]
  10. Wang, Y.-Y.; Sun, Y.-N.; Lin, C.-C.K.; Ju, M.-S. Segmentation of nerve fibers using multi-level gradient watershed and fuzzy systems. Artif. Intell. Med. 2012, 54, 189–200. [Google Scholar] [CrossRef] [PubMed]
  11. Zaimi, A.; Duval, T.; Gasecka, A.; Côté, D.; Stikov, N.; Cohen-Adad, J. AxonSeg: Open source software for axon and myelin segmentation and morphometric analysis. Front. Neuroinform. 2016, 10, 37. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  12. Zaimi, A.; Wabartha, M.; Herman, V.; Antonsanti, P.-L.; Perone, C.S.; Cohen-Adad, J. AxonDeepSeg: Automatic axon and myelin segmentation from microscopy data using convolutional neural networks. Sci. Rep. 2018, 8, 3816. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  13. Ciresan, D.; Giusti, A.; Gambardella, L.; Schmidhuber, J. Deep neural networks segment neuronal membranes in electron microscopy images. Adv. Neural Inf. Process. Syst. 2012, 25, 2843–2851. [Google Scholar]
  14. le Couedic, T.; Caillon, R.; Rossant, F.; Joutel, A.; Urien, H.; Rajani, R.M. Deep-learning based segmentation of challenging myelin sheaths. In Proceedings of the 2020 Tenth International Conference on Image Processing Theory, Tools and Applications (IPTA), Paris, France, 9–12 November 2020; pp. 1–6. [Google Scholar]
  15. Kaji, S.; Kida, S. Overview of image-to-image translation using deep neural networks: Denoising, super-resolution, modality-conversion, and reconstruction in medical imaging. Radiol. Phys. Technol. 2019, 12, 235–248. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  16. Eslami, M.; Tabarestani, S.; Albarqouni, S.; Adeli, E.; Navab, N.; Adjouadi, M. Image-to-images translation for multi-task organ segmentation and bone suppression in chest x-ray radiography. IEEE Trans. Med. Imaging 2020, 39, 2553–2565. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  17. Liu, M.-Y.; Tuzel, O. Coupled generative adversarial networks. In Proceedings of the Advances in Neural Information Processing Systems 29 (NIPS 2016), Barcelona, Spain, 5–10 December 2016; Volume 29. Available online: https://proceedings.neurips.cc/paper/2016/hash/502e4a16930e414107ee22b6198c578f-Abstract.html (accessed on 18 June 2022).
  18. Kuncheva, L.I.; Whitaker, C.J. Measures of diversity in classifier ensembles and their relationship with the ensemble accuracy. Mach. Learn. 2003, 51, 181–207. [Google Scholar] [CrossRef]
  19. Burns, W.A. “Thick Sections: Technique and Applications”, Diagnostic Electron Microscopy; Chapter 4; Trump, B.F., Jones, R.J., Eds.; John Wiley & Sons: New York, NY, USA, 1978. [Google Scholar]
  20. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar]
  21. Wang, G.; Sun, J.; Ma, J.; Xu, K.; Gu, J. Sentiment classification: The contribution of ensemble learning. Decis. Support Syst. 2014, 57, 77–93. [Google Scholar] [CrossRef]
  22. Kato, N.; Matsumoto, M.; Kogawa, M.; Atkins, G.J.; Findlay, D.M.; Fujikawa, T.; Oda, H.; Ogata, M. Critical role of p38 MAPK for regeneration of the sciatic nerve following crush injury in vivo. J. Neuroinflamm. 2013, 10, 915. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  23. Dean, D.C., III; O’Muircheartaigh, J.; Dirks, H.; Travers, B.G.; Adluru, N.; Alexander, A.L.; Deoni, S.C. Mapping an index of the myelin g-ratio in infants using magnetic resonance imaging. Neuroimage 2016, 132, 225–237. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  24. Grüning, P.; Palumbo, A.; Landt, S.K.; Heckmann, L.; Brackhagen, L.; Zille, M.; Mamlouk, A.M. Robust and Markerfree in vitro Axon Segmentation with CNNs. In Proceedings of the International Conference on Wireless Mobile Communication and Healthcare, Online, 18 December 2020; pp. 274–284. [Google Scholar]
  25. Di Scandalea, M.L.; Perone, C.S.; Boudreau, M.; Cohen-Adad, J. Deep active learning for axon-myelin segmentation on histology data. arXiv 2019, arXiv:1907.05143. [Google Scholar]
  26. Tóth, V.; Jayaprakash, N.; Abbas, A.; Khan, A.; Zanos, S.; Zanos, T.P. Single-axon level automatic segmentation and feature extraction from immuhistochemical images of peripheral nerves. In Proceedings of the 2020 42nd Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), Montreal, QC, Canada, 20–24 July 2020. [Google Scholar]
  27. Isaacs, J.; Mallu, S.; Batchelor, M. Modification of commercially available image analysis software for semi-automated qualitative analysis of axon regeneration and myelination in the rat sciatic nerve. J. Neurosci. Methods 2014, 233, 45–49. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Five samples of the original images and their corresponding ground truth labels. Labels contain three classes: axon (white), myelin (gray), and background (black). The size of each sample is 512 × 512 pixels with pixel size of 0.15 × 0.15 µm.
Figure 1. Five samples of the original images and their corresponding ground truth labels. Labels contain three classes: axon (white), myelin (gray), and background (black). The size of each sample is 512 × 512 pixels with pixel size of 0.15 × 0.15 µm.
Electronics 11 02277 g001
Figure 2. Overview of the ensemble algorithm pipeline. During the learning step, the training set is fed into both models. The test images are then fed into both models and the prediction matrices are collected following the Softmax function analysis. The probability matrices are then averaged pixel-wise and the argmax function is used for prediction. Finally, the evaluation metrics are used to assess the performance of the ensemble training model by comparing the ground truth test labels with the prediction matrices.
Figure 2. Overview of the ensemble algorithm pipeline. During the learning step, the training set is fed into both models. The test images are then fed into both models and the prediction matrices are collected following the Softmax function analysis. The probability matrices are then averaged pixel-wise and the argmax function is used for prediction. Finally, the evaluation metrics are used to assess the performance of the ensemble training model by comparing the ground truth test labels with the prediction matrices.
Electronics 11 02277 g002
Figure 3. The axon shapes were assumed to be circular. The calculated areas of the axons and myelinated axons from the segmented image were equated to the area of a circle to find the axonal diameter (Ad) and the myelinated fiber diameter (Fd).
Figure 3. The axon shapes were assumed to be circular. The calculated areas of the axons and myelinated axons from the segmented image were equated to the area of a circle to find the axonal diameter (Ad) and the myelinated fiber diameter (Fd).
Electronics 11 02277 g003
Figure 4. Confusion matrices of the ensemble, ADS, and PIT models.
Figure 4. Confusion matrices of the ensemble, ADS, and PIT models.
Electronics 11 02277 g004
Figure 5. Qualitative representation of the segmentation results of three testing whole images using the ADS, PIT, and ensemble models.
Figure 5. Qualitative representation of the segmentation results of three testing whole images using the ADS, PIT, and ensemble models.
Electronics 11 02277 g005
Figure 6. Curve histograms of the morphological data. (A) the G-ratios, (B) axon diameters, and (C) myelin thicknesses. The plots contain the measurements that are based on the ground truth testing images, ensemble, ADS, and PIT segmented images.
Figure 6. Curve histograms of the morphological data. (A) the G-ratios, (B) axon diameters, and (C) myelin thicknesses. The plots contain the measurements that are based on the ground truth testing images, ensemble, ADS, and PIT segmented images.
Electronics 11 02277 g006
Table 1. Summary of the evaluation metrics calculated from the test images for ADS, PIT, and ensemble models. The results shown are the mean of all of the testing images (24 images).
Table 1. Summary of the evaluation metrics calculated from the test images for ADS, PIT, and ensemble models. The results shown are the mean of all of the testing images (24 images).
ModelPixel-Wise AccuracyAxon Metrics Myelin Metrics
Dice SensitivityPrecision Dice SensitivityPrecision
ADS93.3692.1688.0797.0093.3694.2692.67
PIT90.0087.1190.4484.2386.5282.9090.87
Ensemble95.5194.3193.0396.0094.8995.7294.34
Table 2. Summary of the G-ratios calculated from the ground truth testing images and for the corresponding predicted images of the ADS, PIT, and ensemble models.
Table 2. Summary of the G-ratios calculated from the ground truth testing images and for the corresponding predicted images of the ADS, PIT, and ensemble models.
ImagesGround TruthEnsemble ADSPIT
00.50.5250.5720.572
10.5660.5480.530.608
20.5170.5240.4860.587
30.5420.50.4840.531
Total0.5310.5240.5180.574
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Dweiri, Y.; Al-Zanina, M.; Durand, D. Enhanced Automatic Morphometry of Nerve Histological Sections Using Ensemble Learning. Electronics 2022, 11, 2277. https://doi.org/10.3390/electronics11142277

AMA Style

Dweiri Y, Al-Zanina M, Durand D. Enhanced Automatic Morphometry of Nerve Histological Sections Using Ensemble Learning. Electronics. 2022; 11(14):2277. https://doi.org/10.3390/electronics11142277

Chicago/Turabian Style

Dweiri, Yazan, Mousa Al-Zanina, and Dominique Durand. 2022. "Enhanced Automatic Morphometry of Nerve Histological Sections Using Ensemble Learning" Electronics 11, no. 14: 2277. https://doi.org/10.3390/electronics11142277

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop