You are currently viewing a new version of our website. To view the old version click .
International Journal of Molecular Sciences
  • Article
  • Open Access

16 August 2021

Quantification of the Immune Content in Neuroblastoma: Deep Learning and Topological Data Analysis in Digital Pathology

,
,
,
,
and
1
Data Science for Health, Fondazione Bruno Kessler, 38123 Trento, Italy
2
CIBIO Department, University of Trento, 38123 Trento, Italy
3
Department of Paediatric Haematology/Oncology and of Cell and Gene Therapy, Ospedale Pediatrico Bambino Gesù IRCCS, 00146 Rome, Italy
*
Author to whom correspondence should be addressed.
This article belongs to the Special Issue Neuroblastoma Molecular Biology and Therapeutics

Abstract

We introduce here a novel machine learning (ML) framework to address the issue of the quantitative assessment of the immune content in neuroblastoma (NB) specimens. First, the EUNet, a U-Net with an EfficientNet encoder, is trained to detect lymphocytes on tissue digital slides stained with the CD3 T-cell marker. The training set consists of 3782 images extracted from an original collection of 54 whole slide images (WSIs), manually annotated for a total of 73,751 lymphocytes. Resampling strategies, data augmentation, and transfer learning approaches are adopted to warrant reproducibility and to reduce the risk of overfitting and selection bias. Topological data analysis (TDA) is then used to define activation maps from different layers of the neural network at different stages of the training process, described by persistence diagrams (PD) and Betti curves. TDA is further integrated with the uniform manifold approximation and projection (UMAP) dimensionality reduction and the hierarchical density-based spatial clustering of applications with noise (HDBSCAN) algorithm for clustering, by the deep features, the relevant subgroups and structures, across different levels of the neural network. Finally, the recent TwoNN approach is leveraged to study the variation of the intrinsic dimensionality of the U-Net model. As the main task, the proposed pipeline is employed to evaluate the density of lymphocytes over the whole tissue area of the WSIs. The model achieves good results with mean absolute error 3.1 on test set, showing significant agreement between densities estimated by our EUNet model and by trained pathologists, thus indicating the potentialities of a promising new strategy in the quantification of the immune content in NB specimens. Moreover, the UMAP algorithm unveiled interesting patterns compatible with pathological characteristics, also highlighting novel insights into the dynamics of the intrinsic dataset dimensionality at different stages of the training process. All the experiments were run on the Microsoft Azure cloud platform.

1. Introduction

Neuroblastoma (NB) is the most common cancer diagnosed in the first year of life [1], affecting the sympathetic nervous system. NB is a heterogeneous disease with different outcomes, ranging from spontaneous regression to aggressive progression, metastasis and death. Two main staging systems have been created to stratify patients based on the wide range of outcome and tumor biology: the International Neuroblastoma Staging System (INSS), introduced in 1988 [2] and revised in 1993 [3], and the International Neuroblastoma Risk Group Staging System (INRGSS), introduced in 2009 by the International Neuroblastoma Risk Group task force [4]. INRGSS enhances INSS by defining a series of imaging defined risk factors based on radiological data, such as CT scans and MRI, assessing whether the tumor is circumscribed, if it has metastasized or if it develops near vital parts of the body. However, the effort of establishing an accurate staging system is still ongoing: for instance, additional factors, like the MYCN status, the histopathologic classification, and the DNA content have also proven to be significant [5], and they are currently evaluated in clinical practice. Tumor microenvironment (TME) has been shown to play a role in tumor development. In particular, several pieces of evidence highlighted the importance of the composition, density and distribution of tumor-infiltrating lymphocytes as prognostic markers in several human cancers, including NB. Such observations are stimulating a growing research flow, targeting the dynamics of the immune system during the NB evolution [6], following the Precision Medicine paradigm [7,8,9,10].
Here, we propose a novel artificial intelligence (AI) procedure, sketched in Figure 1, for the quantitative evaluation of the immune content in NB specimens in an immunohistochemistry (IHC) / digital pathology (DP) framework, i.e., using whole slide histopathological images (WSIs) as input data: in detail, a deep learning (DL) predictive model is trained to estimate the density of lymphocytes over the whole tissue area of the WSIs. The approach is demonstrated on the neuroblastoma specimens with T-Lymphocytes - Bambino Gesù (NeSTBG), an original dataset of samples from NB patients, provided by Ospedale Pediatrico Bambino Gesù (OPBG) in Rome, achieving a satisfying performance ( MAE 3.1 ). To boost the reproducibility and interpretability of the DL model, the extracted deep features are analyzed by topological data analysis (TDA) methods [11,12] and, in particular, persistent homology (PH) [13]. To date, this is the first realization of an explainable artificial intelligence (XAI) reproducible platform, encompassing all the analysis steps, from WSI preprocessing to clinical feature interpretation, integrating topological concepts with deep architectures. Although currently more an effective proof of concepts than a fully fledged infrastructure, the novel link established between DL and TDA in DP can lead to further developments along this research line.
Figure 1. Graphical representation of the full analysis workflow. From the original WSIs, the collection of tiles constituting the NeSTBG dataset is generated and annotated to obtain the ground truth for the model (Section 4.1). Tiles are then used as the input for the DL architecture EUNet (Section 4.2 and Section 4.3) to predict density maps (Section 4.4), that are then post-processed and analysed via TDA descriptors to interpret the detected deep features (Section 4.5).

3. Results and Discussion

3.1. Quantification of the Immune Content

To quantify the immune content in NB in terms of lymphocyte detection, a suite of DL experiments were run on the NeSTBG dataset, employing a U-Net network with an EfficientNet-b3 architecture as encoder (EUNet for short). The whole dataset was first partitioned into training (TR) and test (TS) subsets, with ratio 3 4 1 4 ; on the TR portion, a 5-fold cross validation was run four times (TR-CV), and the model trained on the whole TR was then evaluated on the left-out TS. The outcome of the prediction on TS was finally postprocessed (TSp) to enhance the lymphocyte detection: for this model, Precision = 0.73, Recall = 0.85 and F 1 −score = 0.75. The complete set of classification performance is summarized in Table 1.
Table 1. Classification results (in terms of mean and standard deviation mean (st.dev) of the performance metrics for repeated experiments) for different EUNet models: in cross-validation on the training set (TR-CV), in training on the whole TR and inference on the test set, before (TS) and after (TSp) postprocessing (see Section 4.4). MCC: Matthews correlation coefficient; K: Cohen Kappa; MAE: Mean absolute error; MSE: Mean-squared error.
The EUNet was later applied to the entire tissue area of the 54 NeSTBG WSIs to obtain a patient-wise estimate of T-cell density. The tiles already included in NeSTBG were discarded during the training phase to avoid data leakage effects. Note that, for each WSI, NeSTBG includes approximately 1 100 of all possible tiles. In Figure 5, the process of density estimation is graphically summarised on two tiles, while in Figure 6, the effect of postprocessing on the same two tiles is shown.
Figure 5. Examples of the density map prediction process on two tiles. First row: original images. Second row: ground truth density maps. Third row: predicted density maps. Fourth row: predicted density maps discretized with Otsu threshold overlaid on the original image.
Figure 6. Examples of lymphocytes detection obtained through postprocessing of predicted density maps on the original tiles of Figure 5. gt: ground truth after postprocessing, pred: prediction.
To compute the density, the area (mm 2 ) of a single tile of size 512 × 512 pixels can be approximated as A tile = l 2 = 0.655 mm 2 where l tile = (512 pixel mm · ρ · 10 3 ) = 0.256 mm is the tile side length and ρ = 0.5 µ m pixel is the resolution (20×) used for the tile extraction.
As a benchmark, the DL estimate is compared with the manual estimate of a pathologist through the formula proposed in the reference work [22], expressing the density estimate L for each slide as the natural logarithm of the number of lymphocytes per mm 2 :
L = log 1 n i = 0 n c i A i ,
where n is the number of regions of interest (5 or 10), log is the natural logarithm, c i is the number of lymphocytes in the i-th selected region of interest, and A i is the area of the i-th region of interest expressed in mm 2 . The two density estimates have a Pearson correlation coefficient of 0.47 with p-value 3 × 10 4 : in detail, in Figure 7 the corresponding correlation plot is shown, together with the residual plot displaying the difference between DL-predicted density value and pathologist estimation, indicating a positive offset.
Figure 7. Comparison between DL and human expert density estimations computed for the 54 patients of the NeSTBG dataset. Notice that the pathologist estimation is computed on 10 regions of interest while the DL predicted densities are computed on the whole slide. Left panel: correlation plot for predicted densities and density estimated by pathologists. Right panel: residual plot for difference between DL and pathologist density estimation. Yellow line, both panels: perfect correlation. Red line: average difference between DL and pathologist prediction.

3.2. Clinical Assessment of the Topological Features

Clustering analysis was performed by HDBSCAN on the deep features projected by UMAP from the deepest (central) layers of the EUNet. Notably, these features are represented by vectors v i R D , with dimension D = 524,288, as the output of the feature maps in the deepest layers has spatial dimensions 128 × 128 and 64 feature channels. The most interesting structure emerged in the second block of the EUNet decoder; Figure 8 shows the cluster assignment using cosine similarity as metric in the higher dimensional space, 15 neighbours and zero minimum distance for UMAP, minimum cluster size 5 and minimum number of samples 16 for HDBSCAN.
Figure 8. UMAP 2-dimensional embedding of a deep EUNet layer with cosine metric, and HDBSCAN clustering assignments. Gray points (label −1) are classified as noise by the clustering algorithm, while colored points belong to the clusters 0–4 detailed in Figure 9, Figure 10, Figure 11, Figure 12 and Figure 13: stroma rich areas with low TILS level (0), tissue with septa infiltration (1), tissue with pseudo-necrotic tissue infiltration (2), intermediate level of lymphocyte infiltration in stroma poor areas (3) and low level of infiltration in stroma poor areas (4). In the upper-left corner, the graphical schema of the corresponding layer in EUNet.
The tiles belonging to the 5 clusters identified by HDBSCAN can be clinically characterized according to their spatial arrangement. In particular,
  • In cluster 0 (blue), the majority of tiles represents stroma rich areas with low level of TILs (Figure 9).
    Figure 9. Example of 8 representative tiles with stroma rich areas with low level of TILs, grouped as cluster 0 by UMAP and HDBSCAN.
  • In cluster 1 (orange), the majority of the tiles represents tissue with infiltration inside septa (Figure 10).
    Figure 10. Example of 8 representative tiles with infiltration inside septa, grouped as cluster 1 by UMAP and HDBSCAN.
  • In cluster 2 (green), the corresponding tiles present infiltration of lymphocytes in pseudo-necrotic tissue (Figure 11).
    Figure 11. Example of 8 representative tiles with infiltration of lymphocytes in pseudo-necrotic tissue, grouped as cluster 2 by UMAP and HDBSCAN.
  • In cluster 3 (red), the corresponding tiles show an intermediate level of lymphocyte infiltration in stroma poor areas (Figure 12).
    Figure 12. Example of 8 representative tiles with an intermediate level of lymphocyte infiltration in stroma poor areas, grouped as cluster 3 by UMAP and HDBSCAN.
  • In cluster 4 (purple), the corresponding tiles display a low level of infiltration in stroma poor areas (Figure 13).
    Figure 13. Example of 8 representative tiles with a low level of infiltration in stroma poor areas, grouped as cluster 4 by UMAP and HDBSCAN.
The cosine metric seems to be more effective in detecting sub-structures among samples described by DL features than alternative distances such as L 1 or L 2 , as shown in Figure 14.
Figure 14. Alternative UMAP 2-dimensional embedding on the features extracted from the second layer of the decoder path of the EUNet network, with L 1 (a) and L 2 (b) metric. Colors correspond to the detected clusters.
Here, sparsity plays a crucial role: data represent activation maps returned by a rectifier linear unit (ReLU) layer [100] inducing sparsity on the data. Indeed, the extracted feature vectors are quite sparse, with about 60 % of the entries being zero, on average. Given the high-dimensionality and the sparsity, cosine similarity is more effective than L p alternatives [101].
Nonetheless, an interesting pattern emerges also from the UMAP projection of the second layer of the decoder using the Euclidean distance, shown in the two panels of Figure 15. In the left panel, colors represent the INRGS stage of the NB patients, while in the right panel, NB patients are represented according to their MYCN amplification status. In the left scatterplot, high-risk NB patients from stage M are mostly localized on the left portion of the point cloud. Tiles from patients in the L1 stage can be mainly found along the sides of the triangular shape and, finally, most of the tiles from MS patients (with metastases but with favorable prognosis) lie in the centers spreading to the upper and bottom-right vertex of the triangle. Notably, patients with MYCN amplification are clustered together in the upper-left area of the scatter cloud, similarly to high-risk NB patients.
Figure 15. UMAP 2-dimensional embedding on the second layer of the decoder path of the EUNet, with Euclidean metric, minimum distance 0.02 and 15 neighbours. (a) Color indicates INRGSS. The red and green ovals mark the plot areas enclosing the majority of NB patients of stage MS and M, respectively. (b) Color indicates MYCN amplification.

3.3. Topological Analysis of the Deep Features

We computed persistence diagrams (PD) to extract Betti curves from six selected EUNet blocks at different stages during model training. In Figure 16, Betti curves are shown for the 0-th homology group H0 from the third decoder block at different epochs (left panel), with a focus on first three and last three epochs (right panel).
Figure 16. Betti curves for the 0-th homology group H0 from the third decoder block (inset) at different epochs (a) and in particular for the first and last three epochs (b).
Notice that the Betti curves become smoother as the training proceeds, suggesting that the EUNet is progressively learning a meaningful representation of the data. At earlier training stages, several groups of connected components are merged together at uniformly spaced thresholds; later in the process, the curves decrease slower, implying that, from a set of points lying at uniform distances, there are larger groups at non-uniform mutual distances. Finally, towards the end of the training, points become less and less uniformly distributed, as indicated by the smoother profiles of the curves.

3.4. Intrinsic Dimensionality of Datasets

The intrinsic dimensionality (ID) of NeSTBG is computed by the TwoNN algorithm in the six inner blocks of the EUNet (Figure 17) at different stages of the training process. Despite the high dimensionality of the deep feature space, the NeSTBG dataset possibly lives on a manifold of much lower dimension, similarly to the findings in [98]. Specifically, we computed the activation map from the EUNet model state every six epochs, and we estimated the dataset ID. Notably, ID = 125 for the original dataset (computed on 20 random tiles extracted from each patient), while ID = 26 for the predicted density map.
Figure 17. Graphical representation of the EUNet architecture. In the red encircling, the 6 inner blocks computing the Intrinsic Dimensionality at different stages of the training process using the TwoNN algorithm.
Detailed dynamics of the ID estimates are reported in Figure 18. In the top panel, the ID is plotted for each inner block for all the training epochs. In (panel b), ID is plotted for the first three epochs (1,7,13) and for the last epoch (60), which corresponds to the highest peak of the encoder. During the central epochs (panel c) ID values of the encoder are stable, while the ID values of the decoder still show some variability; in particular (panel d) a ID peak on the third block. Thus, ID dynamics share a similar trend in both the encoding and the decoding phase, at different magnitudes.
Figure 18. Intrinsic dimensionality (ID) along different layers of the EUNet network, at different training stages. (a) ID across layers for all training checkpoints, (b) ID across layers for first three training checkpoints, (c) ID across layers for intermediate training checkpoints, (d) ID across layers for intermediate checkpoints, zoomed on the decoder. Legend includes all different epochs; ID curves corresponding to transparent elements of the legend are not shown.

4. Materials and Methods

4.1. The NeSTBG Dataset

The NeSTBG dataset is a collection of 3782 tiles with annotations for the centers of lymphocytes for 54 IHC-stained WSIs of as many NB specimens, previously characterized for density of tumor infiltrating immune cells, including T cells [22], dendritic cells and natural killer cells [102], as well as the expression of PD-L1 and PD-1 [103]. CD3 stained slides were digitalized by the Menarini D-Sight scanner at native magnification 40× (resolution 0.25 µ m pixel ) and employed for digital analysis. The 54 patients are reasonably gender balanced (30 males vs. 24 females), mostly younger than 4 years at diagnosis. INSS, INRGSS and COG values are quite heterogeneous, as well as the tumour location, with suprarenal position as the most frequent (24 patients, 44%); less frequent locations include lymph nodes, aorta, scapula, eye, pharynx, and spleen. The full set of clinical characteristics of the 54 patients are summarised in Figure 19. Morphologically, the large majority of the tumours in the cohort are stroma poor (91%), and in particular poorly differentiated (42 patients, 78%). The remaining 12 samples include 4 differentiated and 3 undifferentiated stroma poor cases, together with an undifferentiated case and 4 ganglioneuroblastoma (GNBL), with only a single stroma rich case. Furthermore, at a 40× magnification level, all samples have about 560 tumoral cells in each sector, while pseudonecrosis areas are mainly present in Stage 4 samples. Note that the heterogeneity of the stroma in the cohort does not represent a confounding factor in the analysis: our experience suggests that immune cells can infiltrate the tumour tissue regardless of the morphology of the stroma, thus yielding that tissue composition is not directly correlated to the immune content. Furthermore, CD3 staining is extremely clean and specific, and the background noise is reduced by precise stain tuning and by blocking the non-specific binding sites, with no need for preprocessing procedures reducing stain variability. Each tile in NeSTBG is a 512 by 512 pixels RGB image stored in png format, randomly extracted from a WSI at 20× magnification.
Figure 19. Summary of clinical features and age distribution at diagnosis (month) for the 54 patients of the NeSTBG dataset.
Annotations refer to the x and y coordinates of the centers of the lymphocytes found in each tile. Level 1 in the OpenSlide standard [104], corresponding to 20× magnification and 0.5 µ m pixel resolution, was selected for the images as a trade-off between image details and computational load, being sufficiently detailed to detect marked cells and to describe WSIs using a limited number of tiles. Segmentation of the tissue region within the slide was also needed: a large portion of WSI is background, and restricting computations only on the tissue area saves both time and resources. However, the original slides in the NeSTBG dataset included many types of artifacts, for instance different appearances of the background surrounding the tissue; WSIs presented a wide range of shades, from pure white to greys with different level of details.
To address the above issues, a sequence of filters were applied to mask out low frequency areas, and morphological operations were used to refine the result. The extraction scheme was designed by overlaying a grid on the tissue area detected on each slide, where each cell of the grid represented a tile. A random number of tiles ranging from 20 to 175 were extracted with random uniform probability, in order to have a representative sample of tissue per slide. The pre-processing steps have been performed with the histolab library (https://github.com/histolab/histolab) (accessed on 13 August 2021), a recently introduced open source Python package for reproducible preprocessing in DP. An example of the tile extraction procedure is shown in Figure 20.
Figure 20. Visualization of the random extraction pattern for the tile extraction in a grid-like fashion. A portion of an CD3+ stained WSI used for the NeSTBG dataset is portrayed (at magnification 1.25×). The size of each tile is representative of the real portion of tissue captured with a 512 × 512 tile at 20× magnification.
The point-wise annotations for the centers of the lymphocytes were manually performed using the VIA annotation tool (version 2) [105] by four trained annotators, generating 73,571 annotations for 3782 tiles extracted from a total of 54 WSIs. Examples of annotations are reported in Figure 21.
Figure 21. Example tiles from NeSTBG with corresponding manual point-wise annotations for the centers of the lymphocytes by the VIA software.
Given the non-negligible irregularity in the shapes of lymphocytes, the staining variability, and the presence of really packed clusters of T-cells, a relaxed constraint for the annotations was chosen, following the strategy introduced in [68] for object counting in crowded scenes; the authors defined a density map of the objects in a crowded scene by centering at each annotated point a Gaussian curve, and normalizing such that the integral over the whole scene would result in the number of objects. When used for lymphocyte detection, the density map approach associates to the annotated centers the highest confidence of objectness [106], a measure that decreases with radial distance from the center.
The point-like annotations were used to build targets to train the deep learning model to reproduce the density maps instead of bounding boxes typically used in an object detection task. This approach allows the model to encode the confidence of the annotation during the training phase, and also to leverage the surrounding context for the prediction. To define a density map, let T be an RGB tile of shape ( N × N × 3 ), and A its set of annotations
A = { c k = ( x k , y k ) | k [ 0 , n ] , x k [ 0 , N 1 ] , y k [ 0 , N 1 ] for n N , n }
The density map is then computed as following:
  • Assign a value d to each annotated pixel and define M ^ as:
    M ^ ( i , j ) = d if ( j , i ) = c k f o r k [ 0 , n ] 0 otherwise
  • Define a Gaussian kernel
    G ( x , y ; σ ) = 1 2 π σ 2 e x 2 + y 2 2 σ 2
    and a squared structuring element G K , with side length l < < N and values given by G centered on the midpoint of G K ;
  • Convolve M ^ with G K to obtain the target density map M = M ^ G K .

4.2. EUNet Architecture

EUNet, the chosen architecture for the predictive model, is based on the fully convolutional U-Net [107] in its encoder-decoder version. The aim of the encoder is to extract feature maps at different depth; the corresponding decoder blocks will up-sample feature maps from preceding layers and use feature maps of the encoder to refine the spatial information. Specifically, for each layer of the decoder:
  • The feature map from the preceding layer is up-sampled with standard up-sampling operations, without any trainable parameter.
  • The up-sampled feature map is concatenated with the feature map from the symmetric level of the encoder path on the depth dimension (i.e., adding more feature channels).
  • The concatenated feature map is fed to convolution operations to refine the spatial information and reduce the number of feature channels.
In this work, we leveraged the PyTorch [108] U-Net implementation provided by Yakubovskiy in [109], which includes several encoder architectures and provides pretrained ImageNet weights [110,111,112].
EfficientNet [113] ( b 3 version) was chosen as encoder; moreover, the spatial and channel squeeze and excitation blocks (scSE) [114] were also introduced in the decoder to improve model performance. The proposed framework is illustrated in Figure 22 and includes:
Figure 22. The full EUNet analysis pipeline.
  • encoder and decoder each composed of five blocks;
  • scSE blocks at the end of each decoder block;
  • Decoder blocks with output feature channels of size: 256, 128, 64, 32, 16;
  • Identity function as activation map in the output layer.
The squeeze and excitation blocks (SE), originally introduced in [115], implement a self-attention mechanism to make the network focus on the most relevant feature channels, by first squeezing the spatial dimensions, and then using global information on feature channels to learn a vector of coefficients used as weights for each channel in the input feature map. See Figure 2 in [115] for a graphical scheme of the SE block. In particular, SEs exploit the global average pooling to resize the input feature map M C x H x W to a vector z C x 1 x 1 , where C is the number of feature channels, H and W the height and the width, respectively. The vector z C × 1 × 1 is processed through a pipeline including a linear layer that halves its size, a ReLU activation layer, a second linear layer that recovers the original number of channels C, and finally a sigmoid activation feeding the vector of weights to the channels of the input feature map. Two different versions of SE were later introduced in [114,116], aimed at improving segmentation models by introducing spatial attention components. The former, named squeeze and excitation, works by first learning a mapping that reduces the number of channels in the input feature map from C to 1, hence summarizing information from the C input channels to a single number for each pixel, resulting in a two dimensional feature map. A sigmoid activation function is applied to each pixel of the two-dimensional feature map, providing weights in the range [0,1] for each pixel of the original feature map. The latter is called scSE block and shown in ([114], Figure 1), had the goal of combining the benefits of learning weights for spatial locations and feature channels. The two approaches work in parallel on the input feature map: a 1 × 1 convolution kernel is applied to obtain a two-dimensional one-channel matrix C M , while preserving the spatial dimensions. A pixel-wise sigmoid activation function is then applied to C M , finally obtaining the weight matrix, then multiplied by the input feature matrix on each channel. Two coefficients are obtained for each entry in the input feature map, and choosing their maximum value leads to best results in terms of performance and complexity added to the model [114].
EfficientNets have been introduced in [113], where the authors exploited the network scaling practice, namely, developed a novel baseline network, which can then be scaled up to obtain a more powerful model. Typically, there are 3 main dimensions, along which it is possible to scale a network: depth, width and image resolution. ResNet is a good example of the first two cases: depth ranges from basic ResNet-18 with 18 layers to architectures with 1000+ layers, while width scaling allows reaching the same accuracy as very deep ResNets with reduced training time [117]. Scaling the third dimension, image resolution, is based on the idea that better resolution of input images implies learning patterns that were not discernible at low-res; however, there is a fundamental technical limit in the memory available on the machine used for training. EfficientNets, based on MnasNet [118], implement a novel strategy—called compound scaling—for scaling base neural network architectures by depth, width and resolution, together using a set of coefficients for each dimension. Compound scaling has been validated also on common ResNet architectures and MobileNets, and can improve network performances, provided the existence of a strong baseline model. The available architectures range from the EfficientNet-b0 to the biggest EfficientNet-b7, achieving top performance on ImageNet with many fewer parameters, thus improving in efficiency. In particular, EfficientNet-b3 has 12 × 10 6 parameters [113] and, tested on ImageNet for a 1000-class classification task, EfficientNet-b3 scores 81.6 % in top-1 accuracy, computed as the comparison between the ground truth and the most confident prediction of the model. Furthermore, because of compound scaling, EfficientNets models support interpretability, since they focus on relevant regions when making predictions, as verified by the Class Activation Map [119]. Therefore, using EfficientNets as the encoder in a U-net architecture, allows the decoder to take advantage of the improved spatial attention mechanism of the encoder, and ultimately to improve the reconstruction of high-resolution density maps.

4.3. EUNet Training and Evaluation

The lymphocyte counting task was censored as a classification task, by manually defining classes of lymphocyte density. The density classes used can be represented by the set C = { 0 , 1 , 2 , 3 , 4 , 5 , 6 } , as shown in Table 2.
Table 2. Lymphocyte count binning in ordinal classes.
Let D be a dataset represented by a collection of n tiles: then, γ N n is the vector of ground truth class for the target lymphocytes in each tile and γ ^ R n is the vector containing class predictions. As model performance metrics we used the mean absolute error (MAE), the mean-squared error (MSE), the accuracy (ACC), the Cohen’s Kappa and the Matthews correlation coefficient (MCC). MAE and MSE are the L 1 and L 2 averaged difference between predicted counts and ground truth counts, respectively, while ACC is the averaged matching between the predicted class and the ground truth class.
MAE = i = 1 n | γ i γ ^ i | n MSE = i = 1 n ( γ i γ ^ i ) 2 n ACC = i = 1 n δ γ i , γ ^ i n ,
where δ is the Kronecker delta δ γ i , γ ^ i = 1 if γ i = γ ^ i 0 otherwise .
The Cohen Kappa K [120] is a statistic measure used to evaluate agreement between two classifier, and it is defined as: K = ACC p e 1 p e , where p e is the sum of the probabilities of the two classifiers agreeing on each class by chance. K takes values in [−1,1] where 1 means perfect agreement between classifiers and 0 or lower values mean that the two classifier are agreeing by chance. In this work, K has quadratic weights for non agreeing values, thus attributing less importance to errors among nearby classes, in accordance with our classes having ordinal values.
Matthews correlation coefficient (MCC) is useful to evaluate classification performance when classes are imbalanced [121]. MCC ranges in [−1,1], where 1 and −1 mean perfect classification and complete misclassification respectively, while 0 indicates random predictions. MCC’s multiclass formula reads as follows [122,123]:
MCC = k , l , m = 1 N ( C k k C m l C l k C k m ) k = 1 N [ l = 1 N C l k f , g = 1 f k N C g f ] k = 1 N [ l = 1 N C k l f , g = 1 f k N C f g ] ,
where C a , b is the number of elements in class a incorrectly predicted in class b.
The loss function used for training is the MSE between the ground truth and the predicted density map, computed pixel-wise. Since the L 2 metric penalizes large differences between pixels according to their magnitude, the larger the values of the peaks in the constructed density maps, the higher the relevance: as a result, pixels in proximity of the lymphocyte centers (where the peaks are located) are more easily predicted than pixels of lymphocyte boundaries. Coupling the Gaussian kernel density maps with the MSE loss drives the network to focus on lymphocytes centers using the context in close proximity, but without great penalty for the exact margin reconstruction.
Hyperparameter optimization is done by the Ranger algorithm [124,125], combining the Lookahead procedure [126], and the Radam stabilization strategy [127]. The rectification strategy of [127] works by tuning the variance parameters of adaptive learning rate optimizers (e.g., Adam [128]) for the first iterations, until variance stabilizes with data from a sufficient number of iterations, thus avoiding the optimizer to remain stuck in local minima. The Lookahead strategy [126] improves parameter exploration speed and stability by using two sets of weights for the optimizer. One set of weights is used for fast exploration of the loss landscape, the other set of weights updates with smaller speed and serves as a stabilizing mechanism if the state of the the optimizer get stuck in unwanted local minima of the loss function. Overall, Ranger proved to be more robust and fast with respect to Adam, warranting a stable optimization, providing a high optimal learning rate η = 10 2 , resulting in a faster training phase, especially for the ResNet50, whose training could be reduced from more than 300 epochs to about 80 epochs.
Networks were initialized by using pretrained weights from ImageNet [110,111,112]; alternative strategies such as using weights from fine-tuned ResNet50 pretrained on the public DP dataset Lysto (https://lysto.grand-challenge.org/, accessed on 13 August 2021), did not lead to a significant performance improvement.
To guarantee robustness and reproducibility to the modeling, a preliminary training/test split with ratio 3 4 1 4 was operated and on the training set a 4 × 5 —cross validation resampling strategy was employed, following the guidelines recommended by the US-FDS in their MAQC/SEQC initiatives [129,130]. Metrics are reported indicating average and standard deviation. Moreover, throughout the model training a particular care has been devoted into avoiding overfitting effects such as data (or information) leakage [35]: tiles extracted from the same WSI were not distributed in different training/test data subsets, a careful approach which is now becoming standard in the most recent works being published [131]. Finally, we adopted a plateau learning rate scheduler acted by monitoring metrics on validation set and reducing the learning rate if no improvements occurred for at least ten epochs: the new learning rate was computed as η t + 1 = α η t with α = 0.2 .

4.4. Lymphocytes Spatial Identification

The predicted lymphocytes density map is post-processed through a 3-step pipeline in order to refine the coordinates of the lymphocytes’ centers:
  • First, the predicted density map values are corrected by setting to zero all pixels with negative values. Indeed, the model learns to predict near-zero values for pixels not belonging to lymphocytes, but the prediction may tend to zero in both positive and negative direction, and for the prediction to be a valid density map the negative values should be removed.
  • Secondly, Otsu thresholding algorithm [132] is used to find an optimal value to discretize the density maps in two levels: lymphocytes and background. The Otsu algorithm is the de facto standard for discriminating foreground and background pixels within an image. In detail, the optimal threshold is identified by minimizing intra-class intensity variance (equivalent to maximizing inter-class variance). Since the Otsu algorithm is the one-dimensional discrete analog of Fisher’s discriminant analysis, this procedure coincides with globally optimizing k-means clustering on the intensity histogram. Pixels with values under the threshold are assigned to the background, while pixels with values over the threshold are assigned to the lymphocyte class.
  • Thirdly, in crowded scenarios, the simple segmentation may still result in connected components including more than one pixel. To split connected components on the Otsu mask, the Watershed segmentation algorithm [133] is used to effectively separate a dense single connected component into multiple sub-components. The result of the Watershed technique is a matrix with n connected components with different labels.
Finally, in order to obtain the coordinates of the center, for each connected component in the mask, the coordinates of the center of mass are computed and used as a proxy for the coordinates of the predicted lymphocytes. The goodness of the detection is evaluated by the three metrics Precision, Recall and F1-score, using as input the two sets of points T and P, defined for each tile as:
T = { t k = ( x k , y k ) | k [ 0 , n 1 ] , x k [ 0 , N 1 ] , y k [ 0 , N 1 ] f o r n 1 N , n 1 } P = { p k = ( x k , y k ) | k [ 0 , n 2 ] , x k [ 0 , N 1 ] , y k [ 0 , N 1 ] f o r n 2 N , n 2 } ,
corresponding to the ground truth and the predicted center’s coordinates, respectively. The Hungarian algorithm [134] is used to find the best assignment between ground truth points and predicted lymphocyte centers. Since optimal assignment can fail if the matched points are too far away, each possible match is accepted only if the distance between points is lower than a given threshold t, with Θ ( t ) = s l for s l 4   μ m , i.e., the average size of a lymphocyte [50], corresponding to 8 pixels. Accepted matches are labeled as true positive, while unmatched ground truths are considered false negatives and unmatched predictions false positives. The performance measures are defined as follows:
Precision = | True Positives | | True Positives | + | False Positives | Recall = | True Positives | | True Positives | + | False Negatives | F 1 = Precision · Recall Precision + Recall ,
being the F 1 -score the harmonic mean of precision and recall, and thus providing a unique measure to describe the overall goodness of predictions.

4.5. Deep Features Interpretation

To explore the hidden layers of the model, a subset S N of 1080 tiles was extracted from the NeSTBG dataset, 20 tiles for each available WSI, and then analysed by three different methods, namely UMAP-HDBSCAN clustering, TDA representation and TwoNN dimensionality estimation. First, UMAP is used to project data into a low b-dimensional space with b [ 2 , 12 ] , where the upper bound is recommended in [95] for later feeding the projection into the HDBSCAN algorithm without falling into a computationally intractable task. Indeed, feature maps from raw images or from hidden layers of the network can be up to 10 6 dimensions. For instance, in an intermediate step of the trained U-Net the feature map has height and width 128 and 64 feature channels, resulting in a flattened vector of length 128 × 128 × 64 = 1,048,576 elements, for each tile. Estimating densities in 10 6 dimensions with approximately 10 3 data points would not be feasible without the UMAP dimensionality reduction step. Value of b needs to be a trade-off between computational constraints (lower b) and the effort of preserving most of the original structure of the dataset (higher b): for the current tasks, b = 4 was chosen. Furthermore, the UMAP minimum distance parameter was set to zero to let the embedding be free of arranging points close together; the number of neighbors parameter was set to 25 so that at each iteration UMAP is forced to compare tiles from more than one patient, since at most 20 tiles are extracted from the same WSI; finally, L2, L1 and cosine norm were used as the distance in the original feature space.
The obtained 4-dim projection was thus used as the input for HDBSCAN to extract the dense regions of the embeddings; the clustering was subsequently visualized using a different 2-dim UMAP projection for a qualitative analysis of its global structure.
Next, Betti curves are used to highlight the topological dynamics of the deep features and finally the estimate of the intrinsic dimensionality of S N as a point cloud is provided by the TwoNN algorithm.
Ansuini et al. [98] experimented standard convolutional neural network architectures for classification tasks (VGG, AlexNet, ResNet) and observed a characteristic pattern of intrinsic dimensionality of the deep features along layers in a well trained model. However, EUNet is more complex, with connections across multiple layers and two main branches with inherently different behaviors (encoding and decoding).

5. Conclusions

WSI data from DP are leveraged here to design a human-in-the-loop ML framework that could aid clinicians in NB risk assessment. As a major novelty in the pipeline, cloud computing is used to train a DL model with state-of-the-art architectures to predict density maps, an approach rarely found in DP for IHC-stained specimens. The predictive model is trained on the task of counting lymphocytes, while a post processing pipeline able to detect nuclei is implemented from the predicted density maps, with results aligned with pathologist’s estimates.
Furthermore, novel TDA approaches are employed to study the hidden representation of data as processed by the network. As future developments, different strategy for data augmentation (such as elastic transformations) or different techniques to construct the predicted target density maps can be explored, as well as possible optimization of the model architecture, and different activation and loss functions. Moreover, the current work focused on the CD3 T-cell marker as a proof of principle that can be extended to other immune cell markers to gain a deeper understanding of the role played by the immune system on NB progression.
Finally, the ML framework would strongly benefit from the ability to simultaneously recognise the tumour regions where lymphocytes are localized, e.g., septa, or tumoral nests, and to observe tiles within a larger portion of the slides, in order to gain a higher level of information.
Overall, the promising results emerging from the the current study pave the way towards the development of an effective learning tool aimed at timely and precisely quantifying the immune content within tumoral cells. Building on the awareness raising from the experience gained by previously published works [22,102,103], such a tool can work as a precious support for the pathologist, with an effective impact on the daily routine in clinical setting.
As a future development, we plan to complement the current methodological work by deepening the reported analysis through the study of the contribute of additional markers such as PD-1 and PDL-1, investigating their correlation level with both cell infiltration and patients outcome to strengthen the derived biological insights on NB.

Author Contributions

Conceptualization, N.B., B.P., D.F. and G.J.; methodology, B.P., N.B.; software, B.P. and N.B.; validation, O.M., A.C., D.F., B.P. and N.B.; formal analysis, investigation, resources, data curation O.M., A.C., D.F., B.P. and N.B.; writing—original draft preparation, B.P. and G.J.; writing—review and editing, B.P., N.B. and G.J.; visualization, B.P.; supervision, D.F. and G.J.; project administration, N.B. and G.J.; funding acquisition, G.J. and D.F. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by grants from the Associazione Italiana Ricerca sul Cancro (AIRC) IG 18495 and AIRC IG 24345.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The authors declare that all data supporting the findings of this study are available within the paper. Any other relevant data and code are available from the corresponding author upon reasonable request. The source code is written in Python/PyTorch as a deep learning framework, and it is available at the GitHub repository https://github.com/bru08/ly_decount (accessed on 13 August 2021).

Acknowledgments

The authors are grateful to Alessia Marcolini for her inspiring passion and helpful advice and to Chiara Riccardi and Alessandra Spilimbergo for their help with the WSI annotation.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
ACCAccuracy
AIArtificial Intelligence
ANNArtificial Neural Network
CD3Cluster of Differentiation 3
CNNConvolutional Neural Networks
CTComputer Tomography
DLDeep Learning
DPDigital Pathology
EUNetU-Net with EfficientNet-b3 encoder
GBGigaByte
HDBSCANHierarchical Density-Based Spatial Clustering of Applications with Noise
IHCImmunohistochemistry
INRGSSInternational Neuroblastoma Risk Group Staging System
INSSInternational Neuroblastoma Staging System
MAEMean Absolute Error
MBMegaByte
MCCMatthews Correlation Coefficient
MLMachine Learning
MRIMagnetic Resonance Imaging
MSEMean Squared Error
MYCNv-myc avian myelocytomatosis viral oncogene neuroblastoma derived homolog
NeSTBGNeuroblastoma Specimens with T-Lymphocytes - Bambino Gesù (dataset)
OPBGOspedale Pediatrico Bambin Gesù
NBNeuroblastoma
PDPersistent Diagram
PHPersistent Homology
ResNetResidual Neural Networks
RGBRed Green Blue
scSEspatial and channel Squeeze & Excitation (block)
SESqueeze & Excitation (block)
TDATopological Data Analysis
TILTumor Inflitrating Lymphocytes
UMAPUniform Manifold Approximation and Projection
WSIWhole Slide Image

References

  1. Cheung, N.K.V.; Dyer, M.A. Neuroblastoma: Developmental biology, cancer genomics and immunotherapy. Nat. Rev. Cancer 2013, 13, 397–411. [Google Scholar] [CrossRef] [Green Version]
  2. Brodeur, G.M.; Seeger, R.C.; Barrett, A.; Berthold, F.; Castleberry, R.P.; D’Angio, G.; De Bernardi, B.; Evans, A.E.; Favrot, M.; Freeman, A.I. International criteria for diagnosis, staging, and response to treatment in patients with neuroblastoma. J. Clin. Oncol. 1988, 6, 1874–1881. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Brodeur, G.M.; Pritchard, J.; Berthold, F.; Carlsen, N.L.; Castel, V.; Castelberry, R.P.; De Bernardi, B.; Evans, A.E.; Favrot, M.; Hedborg, F. Revisions of the international criteria for neuroblastoma diagnosis, staging, and response to treatment. J. Clin. Oncol. 1993, 11, 1466–1477. [Google Scholar] [CrossRef] [PubMed]
  4. Cohn, S.L.; Pearson, A.D.J.; London, W.B.; Monclair, T.; Ambros, P.F.; Brodeur, G.M.; Faldum, A.; Hero, B.; Iehara, T.; Machin, D.; et al. The International Neuroblastoma Risk Group (INRG) Classification System: An INRG Task Force Report. J. Clin. Oncol. 2009, 27, 289–297. [Google Scholar] [CrossRef] [PubMed]
  5. Davidoff, A.M. Neuroblastoma. Semin. Pediatr. Surg. 2012, 21, 2–14. [Google Scholar] [CrossRef] [Green Version]
  6. Wienke, J.; Dierselhuis, M.P.; Tytgat, G.A.M.; Künkele, A.; Nierkens, S.; Molenaar, J.J. The immune landscape of neuroblastoma: Challenges and opportunities for novel therapeutic strategies in pediatric oncology. Eur. J. Cancer 2021, 144, 123–150. [Google Scholar] [CrossRef]
  7. Barsoum, I.; Tawedrous, E.; Faragalla, H.; Yousef, G.M. Histo-genomics: Digital pathology at the forefront of precision medicine. Diagnosis 2019, 6, 203–212. [Google Scholar] [CrossRef] [PubMed]
  8. Filipp, F.V. Opportunities for Artificial Intelligence in Advancing Precision Medicine. Curr. Genet. Med. Rep. 2019, 7, 208–213. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  9. Irwin, M.S.; Park, J.R. Neuroblastoma: Paradigm for precision medicine. Pediatr. Clin. N. Am. 2015, 62, 225–256. [Google Scholar] [CrossRef]
  10. Collins, F.S.; Varmus, H. A new initiative on precision medicine. N. Engl. J. Med. 2015, 372, 793–795. [Google Scholar] [CrossRef] [Green Version]
  11. Carlsson, G. Topological methods for data modelling. Nat. Rev. Phys. 2020, 2, 697–708. [Google Scholar] [CrossRef]
  12. Singh, G.; Memoli, F.; Carlsson, G. Topological Methods for the Analysis of High Dimensional Data Sets and 3D Object Recognition. In Proceedings of the IEEE/Eurographics Symposium on Point-Based Graphics 2007 (PBG), Prague, Czech Republic, 2–3 September 2007; pp. 91–100. [Google Scholar]
  13. Huber, S. Persistent Homology in Data Science. In Data Science—Analytics and Applications—Proc. 3rd International Data Science Conference 2020 (iDSC); Springer Fachmedien Wiesbaden: Berlin/Heidelberg, Germany, 2021; pp. 81–88. [Google Scholar]
  14. Kim, S.W.; Roh, J.; Park, C.S. Immunohistochemistry for pathologists: Protocols, pitfalls, and tips. J. Pathol. Transl. Med. 2016, 50, 411–418. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Stanton, S.E.; Disis, M.L. Clinical significance of tumor-infiltrating lymphocytes in breast cancer. J. Immunother. Cancer 2016, 4, 59. [Google Scholar] [CrossRef] [Green Version]
  16. Fu, Q.; Chen, N.; Ge, C.; Li, R.; Li, Z.; Zeng, B.; Li, C.; Wang, Y.; Xue, Y.; Song, X.; et al. Prognostic value of tumor-infiltrating lymphocytes in melanoma: A systematic review and meta-analysis. OncoImmunology 2019, 8, e1593806. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  17. Clemente, C.G.; Mihm, M.C.; Bufalino, R.; Zurrida, S.; Collini, P.; Cascinelli, N. Prognostic value of tumor infiltrating lymphocytes in the vertical growth phase of primary cutaneous melanoma. Cancer Interdiscip. Int. J. Am. Cancer Soc. 1996, 77, 1303–1310. [Google Scholar] [CrossRef]
  18. Martin, R.F.; Beckwith, J.B. Lymphoid infiltrates in neuroblastomas: Their occurrence and prognostic significance. J. Pediatr. Surg. 1968, 3, 161–164. [Google Scholar] [CrossRef]
  19. Lauder, I.; Aherne, W. The Significance of Lymphocytic Infiltration in Neuroblastoma. Br. J. Cancer 1972, 26, 321–330. [Google Scholar] [CrossRef] [Green Version]
  20. Facchetti, P.; Prigione, I.; Ghiotto, F.; Tasso, P.; Garaventa, A.; Pistoia, V. Functional and molecular characterization of tumour-infiltrating lymphocytes and clones thereof from a major-histocompatibility-complex-negative human tumour: Neuroblastoma. Cancer Immunol. Immunother. 1996, 42, 170–178. [Google Scholar] [CrossRef]
  21. Carlson, L.M.; De Geer, A.; Sveinbjørnsson, B.; Orrego, A.; Martinsson, T.; Kogner, P.; Levitskaya, J. The microenvironment of human neuroblastoma supports the activation of tumor-associated T lymphocytes. OncoImmunology 2013, 2, e23618. [Google Scholar] [CrossRef] [Green Version]
  22. Mina, M.; Boldrini, R.; Citti, A.; Romania, P.; D’Alicandro, V.; De Ioris, M.; Castellano, A.; Furlanello, C.; Locatelli, F.; Fruci, D. Tumor-infiltrating T lymphocytes improve clinical outcome of therapy-resistant neuroblastoma. OncoImmunology 2015, 4, e1019981. [Google Scholar] [CrossRef] [Green Version]
  23. Kaplan, K.J.; Rao, L.K.F. Digital Pathology: Historical Perspectives, Current Concepts & Future Applications; Springer: Berlin/Heidelberg, Germany, 2015. [Google Scholar]
  24. Zarella, M.D.; Bowman, D.; Aeffner, F.; Farahani, N.; Xthona, A.; Absar, S.F.; Parwani, A.; Bui, M.; Hartman, D.J. A Practical Guide to Whole Slide Imaging: A White Paper From the Digital Pathology Association. Arch. Pathol. Lab. Med. 2018, 143, 222–234. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  25. Echle, A.; Rindtorff, N.T.; Brinker, T.J.; Luedde, T.; Pearson, A.T.; Kather, J.N. Deep learning in cancer pathology: A new generation of clinical biomarkers. Br. J. Cancer 2021, 124, 686–696. [Google Scholar] [CrossRef] [PubMed]
  26. Barisoni, L.; Lafata, K.J.; Hewitt, S.M.; Madabhushi, A.; Balis, U.G.J. Digital pathology and computational image analysis in nephropathology. Nat. Rev. Nephrol. 2020, 16, 669–685. [Google Scholar] [CrossRef] [PubMed]
  27. Hägele, M.; Seegerer, P.; Lapuschkin, S.; Bockmayr, M.; Samek, W.; Klauschen, F.; Müller, K.R.; Binder, A. Resolving challenges in deep learning-based analyses of histopathological images using explanation methods. Sci. Rep. 2020, 10, 6423. [Google Scholar] [CrossRef] [Green Version]
  28. Iizuka, O.; Kanavati, F.; Kato, K.; Rambeau, M.; Arihiro, K.; Tsuneki, M. Deep Learning Models for Histopathological Classification of Gastric and Colonic Epithelial Tumours. Sci. Rep. 2020, 10, 1504. [Google Scholar] [CrossRef] [Green Version]
  29. Song, Z.; Zou, S.; Zhou, W.; Huang, Y.; Shao, L.; Yuan, J.; Gou, X.; Jin, W.; Wang, Z.; Chen, X.; et al. Clinically applicable histopathological diagnosis system for gastric cancer detection using deep learning. Nat. Commun. 2020, 11, 4294. [Google Scholar] [CrossRef]
  30. Acs, B.; Rantalainen, M.; Hartman, J. Artificial intelligence as the next step towards precision pathology. J. Intern. Med. 2020, 288, 62–81. [Google Scholar] [CrossRef] [Green Version]
  31. Nagpal, K.; Foote, D.; Liu, Y.; Chen, P.H.C.; Wulczyn, E.; Tan, F.; Olson, N.; Smith, J.L.; Mohtashamian, A.; Wren, J.H.; et al. Development and validation of a deep learning algorithm for improving Gleason scoring of prostate cancer. Npj Digit. Med. 2019, 2, 48. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  32. Wulczyn, E.; Steiner, D.F.; Xu, Z.; Sadhwani, A.; Wang, H.; Flament-Auvigne, I.; Mermel, C.H.; Chen, P.H.C.; Liu, Y.; Stumpe, M.C. Deep learning-based survival prediction for multiple cancer types using histopathology images. PLoS ONE 2020, 15, e0233678. [Google Scholar]
  33. Tellez, D.; Litjens, G.; van der Laak, J.; Ciompi, F. Neural image compression for gigapixel histopathology image analysis. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 43, 567–578. [Google Scholar] [CrossRef] [Green Version]
  34. Brancati, N.; De Pietro, G.; Riccio, D.; Frucci, M. Gigapixel Histopathological Image Analysis using Attention-based Neural Networks. arXiv 2021, arXiv:2101.09992. [Google Scholar]
  35. Bussola, N.; Marcolini, A.; Maggio, V.; Jurman, G.; Furlanello, C. AI slipping on tiles: Data leakage in digital pathology. In Lecture Notes in Computer Science, Proceedings of the International Workshop on Artificial Intelligence for Digital Pathology 2021 (AIDP); Springer International Publishing: Berlin/Heidelberg, Germany, 2021; Volume 12661, pp. 167–182. [Google Scholar]
  36. Varn, F.S.; Wang, Y.; Mullins, D.W.; Fiering, S.; Cheng, C. Systematic Pan-Cancer Analysis Reveals Immune Cell Interactions in the Tumor Microenvironment. Cancer Res. 2017, 77, 1271–1282. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  37. Swiderska-Chadaj, Z.; Pinckaers, H.; van Rijthoven, M.; Balkenhol, M.; Melnikova, M.; Geessink, O.; Manson, Q.; Sherman, M.; Polonia, A.; Parry, J.; et al. Learning to detect lymphocytes in immunohistochemistry with deep learning. Med. Image Anal. 2019, 58, 101547. [Google Scholar] [CrossRef]
  38. Salvi, M.; Acharya, U.R.; Molinari, F.; Meiburger, K.M. The impact of pre- and post-image processing techniques on deep learning frameworks: A comprehensive review for digital pathology image analysis. Comput. Biol. Med. 2021, 128, 104129. [Google Scholar] [CrossRef]
  39. Janowczyk, A.; Madabhushi, A. Deep learning for digital pathology image analysis: A comprehensive tutorial with selected use cases. J. Pathol. Inform. 2016, 7, 29. [Google Scholar] [CrossRef] [PubMed]
  40. Linder, N.; Taylor, J.C.; Colling, R.; Pell, R.; Alveyn, E.; Joseph, J.; Protheroe, A.; Lundin, M.; Lundin, J.; Verrill, C. Deep learning for detecting tumour-infiltrating lymphocytes in testicular germ cell tumours. J. Clin. Pathol. 2018, 72, 157–164. [Google Scholar] [CrossRef]
  41. Saltz, J.; Gupta, R.; Hou, L.; Kurc, T.; Singh, P.; Nguyen, V.; Samaras, D.; Shroyer, K.R.; Zhao, T.; Batiste, R.; et al. Spatial Organization and Molecular Correlation of Tumor-Infiltrating Lymphocytes Using Deep Learning on Pathology Images. Cell Rep. 2018, 23, 181–193.e7. [Google Scholar] [CrossRef] [Green Version]
  42. Banfield, J.D.; Raftery, A.E. Model-Based Gaussian and Non-Gaussian Clustering. Biometrics 1993, 49, 803. [Google Scholar] [CrossRef]
  43. Bidart, R.; Gangeh, M.J.; Peikari, M.; Martel, A.L.; Ghodsi, A.; Salama, S.; Nofech-Mozes, S. Localization and classification of cell nuclei in post-neoadjuvant breast cancer surgical specimen using fully convolutional networks. In Proceedings of the Medical Imaging 2018: Digital Pathology, Houston, TX, USA, 11–12 February 2018; Volume 10581, p. 1058100. [Google Scholar]
  44. Vincent, L. Morphological grayscale reconstruction in image analysis: Applications and efficient algorithms. IEEE Trans. Image Process. 1993, 2, 176–201. [Google Scholar] [CrossRef] [Green Version]
  45. Li, C.T.; Chung, P.C.; Tsai, H.W.; Chow, N.H.; Cheng, K.S. Inflammatory Cells Detection in H&E Staining Histology Images Using Deep Convolutional Neural Network with Distance Transformation. In Communications in Computer and Information Science; Springer: Singapore, 2019; pp. 665–672. [Google Scholar]
  46. Chen, T.; Chefd’hotel, C. Deep Learning Based Automatic Immune Cell Detection for Immunohistochemistry Images. In Proceedings of the Machine Learning in Medical Imaging 2014 (MLMI), Boston, MA, USA, 14 September 2014; Springer International Publishing: Berlin/Heidelberg, Germany, 2014; pp. 17–24. [Google Scholar]
  47. Garcia, E.; Hermoza, R.; Castanon, C.B.; Cano, L.; Castillo, M.; Castanñeda, C. Automatic Lymphocyte Detection on Gastric Cancer IHC Images Using Deep Learning. In Proceedings of the IEEE 30th International Symposium on Computer-Based Medical Systems 2017 (CBMS), Thessaloniki, Greece, 22–24 June 2017; pp. 200–204. [Google Scholar]
  48. Redmon, J.; Farhadi, A. YOLO9000: Better, faster, stronger. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition 2017 (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 6517–6525. [Google Scholar]
  49. Swiderska-Chadaj, Z.; Pinckaers, H.; van Rijthoven, M.; Balkenhol, M.; Melnikova, M.; Geessink, O.; Manson, Q.; Litjens, G.; van der Laak, J.; Ciompi, F. Convolutional Neural Networks for Lymphocyte detection in Immunohistochemically Stained Whole-Slide Images, 2018. In Proceedings of the Poster at Medical Imaging with Deep Learning 2018 (MIDL), Amsterdam, The Netherlands, 4–6 July 2018. [Google Scholar]
  50. Van Rijthoven, M.; Swiderska-Chadaj, Z.; Seeliger, K.; van der Laak, J.; Ciompi, F. You only look on lymphocytes once. In Proceedings of the Conference on Medical Imaging with Deep Learning 2018 (MIDL); 2018; pp. 1–3. Available online: https://openreview.net/forum?id=S10IfW2oz (accessed on 13 August 2021).
  51. Evangeline, K.; Precious, J.; Pazhanivel, N.; Kirubha, A. Automatic Detection and Counting of Lymphocytes from Immunohistochemistry Cancer Images Using Deep Learning. J. Med. Biol. Eng. 2020, 40, 735–747. [Google Scholar] [CrossRef]
  52. Ghahremani, P.; Li, Y.; Kaufman, A.; Vanguri, R.; Greenwald, N.; Angelo, M.; Hollmann, T.J.; Nadeem, S. DeepLIIF: Deep Learning-Inferred Multiplex ImmunoFluorescenc e for IHC Quantification. bioRxiv 2021. [Google Scholar] [CrossRef]
  53. Kapil, A.; Meier, A.; Shumilov, A.; Haneder, S.; Angell, H.; Schmidt, G. Breast cancer patient stratification using domain adaptation based lymphocyte detection in HER2 stained tissue sections. In Proceedings of the Poster at Medical Imaging with Deep Learning 2021 (MIDL), Lübeck, Germany, 7–9 July 2021. [Google Scholar]
  54. Negahbani, F.; Sabzi, R.; Jahromi, B.P.; Firouzabadi, D.; Movahedi, F.; Shirazi, M.K.; Majidi, S.; Dehghanian, A. PathoNet introduced as a deep neural network backend for evaluation of Ki-67 and tumor-infiltrating lymphocytes in breast cancer. Sci. Rep. 2021, 11, 8489. [Google Scholar] [CrossRef]
  55. Hermsen, M.; Volk, V.; Bräsen, J.H.; Geijs, D.J.; Gwinner, W.; Kers, J.; Linmans, J.; Schaadt, N.S.; Schmitz, J.; Steenbergen, E.J.; et al. Quantitative assessment of inflammatory infiltrates in kidney transplant biopsies using multiplex tyramide signal amplification and deep learning. Lab. Investig. 2021, 101, 970–982. [Google Scholar] [CrossRef] [PubMed]
  56. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition 2014 (CVPR), Columbus, OH, USA, 23–28 June 2014; pp. 580–587. [Google Scholar]
  57. Girshick, R. Fast R-CNN. In Proceedings of the IEEE International Conference on Computer Vision 2015 (ICCV), Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
  58. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef] [Green Version]
  59. Zhang, J.; Wang, X.; Ni, G.; Liu, J.; Hao, R.; Liu, L.; Liu, Y.; Du, X.; Xu, F. Fast and accurate automated recognition of the dominant cells from fecal images based on Faster R-CNN. Sci. Rep. 2021, 11, 10361. [Google Scholar] [CrossRef]
  60. Liu, S.; Zhang, Y.; Ju, Y.; Li, Y.; Kang, X.; Yang, X.; Niu, T.; Xing, X.; Lu, Y. Establishment and Clinical Application of an Artificial Intelligence Diagnostic Platform for Identifying Rectal Cancer Tumor Budding. Front. Oncol. 2021, 11, 320. [Google Scholar]
  61. Lo, Y.C.; Chung, I.F.; Guo, S.N.; Wen, M.C.; Juang, C.F. Cycle-consistent GAN-based stain translation of renal pathology images with glomerulus detection application. Appl. Soft Comput. 2021, 98, 106822. [Google Scholar] [CrossRef]
  62. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask R-CNN. In Proceedings of the IEEE International Conference on Computer Vision 2017 (ICCV), Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar]
  63. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask R-CNN. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 42, 386–397. [Google Scholar] [CrossRef]
  64. Durkee, M.S.; Abraham, R.; Ai, J.; Fuhrman, J.D.; Clark, M.R.; Giger, M.L. Comparing Mask R-CNN and U-Net architectures for robust automatic segmentation of immune cells in immunofluorescence images of Lupus Nephritis biopsies. In Proceedings of the Imaging, Manipulation, and Analysis of Biomolecules, Cells, and Tissues 2021 XIX; Leary, J.F., Tarnok, A., Georgakoudi, I., Eds.; SPIE: Bellingham, WA, USA, 2021; Volume 11647, p. 116470X. [Google Scholar]
  65. Fujita, S.; Han, X.H. Cell Detection and Segmentation in Microscopy Images with Improved Mask R-CNN. In Proceedings of the Computer Vision 2020 Workshops (ACCV); Springer: Berlin/Heidelberg, Germany, 2021; pp. 58–70. [Google Scholar]
  66. Wang, S.; Rong, R.; Yang, D.M.; Fujimoto, J.; Yan, S.; Cai, L.; Yang, L.; Luo, D.; Behrens, C.; Parra, E.R.; et al. Computational Staining of Pathology Images to Study the Tumor Microenvironment in Lung Cancer. Cancer Res. 2020, 80, 2056–2066. [Google Scholar] [CrossRef] [Green Version]
  67. Altini, N.; Cascarano, G.D.; Brunetti, A.; De Feudis, I.; Buongiorno, D.; Rossini, M.; Pesce, F.; Gesualdo, L.; Bevilacqua, V. A Deep Learning Instance Segmentation Approach for Global Glomerulosclerosis Assessment in Donor Kidney Biopsies. Electronics 2020, 9, 1768. [Google Scholar] [CrossRef]
  68. Lempitsky, V.; Zisserman, A. Learning To Count Objects in Images. In Proceedings of the Advances in Neural Information Processing Systems 2012 (NIPS); Curran Associates Inc.: Red Hook, NY, USA, 2010; pp. 1324–1332. [Google Scholar]
  69. Zhang, Y.; Zhou, D.; Chen, S.; Gao, S.; Ma, Y. Single-Image Crowd Counting via Multi-Column Convolutional Neural Network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition 2016 (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 589–597. [Google Scholar]
  70. Bahmanyar, R.; Vig, E.; Reinartz, P. MRCNet: Crowd Counting and Density Map Estimation in Aerial and Ground Imagery. arXiv 2019, arXiv:1909.12743. [Google Scholar]
  71. Kong, Y.; Li, H.; Ren, Y.; Genchev, G.Z.; Wang, X.; Zhao, H.; Xie, Z.; Lu, H. Automated yeast cells segmentation and counting using a parallel U-Net based two-stage framework. Osa Contin. 2020, 3, 982–992. [Google Scholar] [CrossRef]
  72. Jiang, N.; Yu, F. Multi-column network for cell counting. Osa Contin. 2020, 3, 1834–1846. [Google Scholar] [CrossRef]
  73. Chazal, F.; Michel, B. An introduction to Topological Data Analysis: Fundamental and practical aspects for data scientists. arXiv 2017, arXiv:1710.04019. [Google Scholar]
  74. Perea, J.A.; Harer, J. Sliding windows and persistence: An application of topological methods to signal analysis. Found. Comput. Math. 2015, 15, 799–838. [Google Scholar] [CrossRef] [Green Version]
  75. Carlsson, G.; Ishkhanov, T.; de Silva, V.; Zomorodian, A. On the Local Behavior of Spaces of Natural Images. Int. J. Comput. Vis. 2007, 76, 1–12. [Google Scholar] [CrossRef]
  76. Rieck, B.; Yates, T.; Bock, C.; Borgwardt, K.; Wolf, G.; Turk-Browne, N.; Krishnaswamy, S. Uncovering the topology of time-varying fMRI data using cubical persistence. In Proceedings of the Advances in Neural Information Processing Systems 2020 (NeurIPS); Curran Associates Inc.: Red Hook, NY, USA, 2020; Volume 33, pp. 6900–6912. [Google Scholar]
  77. Bauer, U. Ripser: Efficient computation of Vietoris-Rips persistence barcodes. arXiv 2019, arXiv:1908.02518. [Google Scholar]
  78. Tauzin, G.; Lupo, U.; Tunstall, L.; Burella Pérez, J.; Caorsi, M.; Medina-Mardones, A.; Dassatti, A.; Hess, K. giotto-tda: A Topological Data Analysis Toolkit for Machine Learning and Data Exploration. arXiv 2020, arXiv:2004.02551. [Google Scholar]
  79. Riihimäki, H.; Chachólski, W.; Theorell, J.; Hillert, J.; Ramanujam, R. A topological data analysis based classification method for multiple measurements. BMC Bioinform. 2020, 21, 336. [Google Scholar] [CrossRef]
  80. Walsh, K.; Voineagu, M.A.; Vafaee, F.; Voineagu, I. TDAview: An online visualization tool for topological data analysis. Bioinformatics 2020, 36, 4805–4809. [Google Scholar] [CrossRef]
  81. Amézquita, E.J.; Quigley, M.Y.; Ophelders, T.; Munch, E.; Chitwood, D.H. The shape of things to come: Topological data analysis and biology, from molecules to organisms. Dev. Dyn. 2020, 249, 816–833. [Google Scholar] [CrossRef] [PubMed]
  82. Mandal, S.; Guzmán-Sáenz, A.; Haiminen, N.; Basu, S.; Parida, L. A Topological Data Analysis Approach on Predicting Phenotypes from Gene Expression Data. In Proceedings of the Algorithms for Computational Biology 2020 (AlCoB); Lecture Notes in Computer Science; Springer International Publishing: Berlin/Heidelberg, Germany, 2020; Volume 12099, pp. 178–187. [Google Scholar]
  83. Shen, C.; Patrangenaru, V. Topological Object Data Analysis Methods with an Application to Medical Imaging. In Proceedings of the Functional and High-Dimensional Statistics and Related Fields 2020 (IWFOS); Springer International Publishing: Berlin/Heidelberg, Germany, 2020; pp. 237–244. [Google Scholar]
  84. Needham, T. Introduction to Applied Algebraic Topology. Course Notes. Available online: https://research.math.osu.edu/tgda/courses/math-4570/LectureNotes.pdf (accessed on 13 August 2021).
  85. Bubenik, P. Statistical topological data analysis using persistence landscapes. J. Mach. Learn. Res. 2015, 16, 77–102. [Google Scholar]
  86. McInnes, L.; Healy, J.; Melville, J. UMAP: Uniform manifold approximation and projection for dimension reduction. arXiv 2018, arXiv:1802.03426. [Google Scholar]
  87. McInnes, L.; Healy, J.; Saul, N.; Grossberger, L. UMAP: Uniform Manifold Approximation and Projection. J. Open Source Softw. 2018, 3, 861. [Google Scholar] [CrossRef]
  88. Schenck, H. Computational Algebraic Geometry; Cambridge University Press: Cambridge, UK, 2003; Volume 58. [Google Scholar]
  89. Carter, S.; Armstrong, Z.; Schubert, L.; Johnson, I.; Olah, C. Exploring Neural Networks with Activation Atlases. 2019. Available online: https://distill.pub/2019/activation-atlas (accessed on 13 August 2021).
  90. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition 2015 (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  91. Sainburg, T.; McInnes, L.; Gentner, T.Q. Parametric UMAP: Learning embeddings with deep neural networks for representation and semi-supervised learning. arXiv 2020, arXiv:2009.12981. [Google Scholar]
  92. Kobak, D.; Linderman, G.C. Initialization is critical for preserving global data structure in both t-SNE and UMAP. Nat. Biotechnol. 2021, 39, 156–157. [Google Scholar] [CrossRef] [PubMed]
  93. Campello, R.J.G.B.; Moulavi, D.; Sander, J. Density-based clustering based on hierarchical density estimates. In Proceedings of the Advances in Knowledge Discovery and Data Mining 2013—Pacific-Asia Conference on Knowledge Discovery and Data Mining (PAKDD); Lecture Notes in Computer Science; Springer International Publishing: Berlin/Heidelberg, Germany, 2013; Volume 7819, pp. 160–172. [Google Scholar]
  94. Ester, M.; Kriegel, H.P.; Sander, J.; Xu, X. A Density-Based Algorithm for Discovering Clusters in Large Spatial Databases with Noise. In Proceedings of the Knowledge Discovery and Data Mining 1996 (KDD), Portland, OR, USA, 2–4 August 1996; Volume 96, pp. 226–231. [Google Scholar]
  95. McInnes, L.; Healy, J.; Astels, S. hdbscan: Hierarchical density based clustering. J. Open Source Softw. 2017, 2, 205. [Google Scholar] [CrossRef]
  96. Odena, A.; Dumoulin, V.; Olah, C. Deconvolution and Checkerboard Artifacts. 2016. Available online: http://distill.pub/2016/deconv-checkerboard (accessed on 13 August 2021).
  97. Gabrielsson, R.B.; Carlsson, G. Exposition and Interpretation of the Topology of Neural Networks. In Proceedings of the IEEE International Conference on Machine Learning and Applications 2019 (ICMLA), Boca Raton, FL, USA, 16–19 December 2019; pp. 1069–1076. [Google Scholar]
  98. Ansuini, A.; Laio, A.; Macke, J.H.; Zoccolan, D. Intrinsic dimension of data representations in deep neural networks. In Proceedings of the Advances in Neural Information Processing Systems 2019 (NeurIPS); Curran Associates, Inc.: Red Hook, NY, USA, 2019; Volume 32, pp. 6111–6122. [Google Scholar]
  99. Facco, E.; d’Errico, M.; Rodriguez, A.; Laio, A. Estimating the intrinsic dimension of datasets by a minimal neighborhood information. Sci. Rep. 2017, 7, 1–8. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  100. Nair, V.; Hinton, G.E. Rectified Linear Units Improve Restricted Boltzmann Machines. In Proceedings of the International Conference on Machine Learning 2010 (ICML); Omnipress: Madison, WI, USA, 2010; pp. 807–814. [Google Scholar]
  101. Sun, Y.; Platoš, J. High-Dimensional Text Clustering by Dimensionality Reduction and Improved Density Peak. Wirel. Commun. Mob. Comput. 2020, 2020, 8881112. [Google Scholar] [CrossRef]
  102. Melaiu, O.; Chierici, M.; Lucarini, V.; Jurman, G.; Conti, L.A.; Vito, R.D.; Boldrini, R.; Cifaldi, L.; Castellano, A.; Furlanello, C.; et al. Cellular and gene signatures of tumor-infiltrating dendritic cells and natural-killer cells predict prognosis of neuroblastoma. Nat. Commun. 2020, 11, 5992. [Google Scholar] [CrossRef]
  103. Melaiu, O.; Mina, M.; Chierici, M.; Boldrini, R.; Jurman, G.; Romania, P.; D’Alicandro, V.; Benedetti, M.C.; Castellano, A.; Liu, T.; et al. PD-L1 Is a Therapeutic Target of the Bromodomain Inhibitor JQ1 and, Combined with HLA Class I, a Promising Prognostic Biomarker in Neuroblastoma. Clin. Cancer Res. 2017, 23, 4462–4472. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  104. Goode, A.; Satyanarayanan, M.; Gilbert, B.; Harkes, J.; Jukic, D. OpenSlide: A vendor-neutral software foundation for digital pathology. J. Pathol. Inform. 2013, 4, 27. [Google Scholar]
  105. Dutta, A.; Zisserman, A. The VIA Annotation Software for Images, Audio and Video. In Proceedings of the ACM International Conference on Multimedia 2019 (MM), Nice, France, 21–25 October 2019; pp. 2276–2279. [Google Scholar]
  106. Alexe, B.; Deselaers, T.; Ferrari, V. Measuring the Objectness of Image Windows. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 34, 2189–2202. [Google Scholar] [CrossRef] [Green Version]
  107. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention 2015 (MICCAI); Lecture Notes in Computer Science; Springer International Publishing: Berlin/Heidelberg, Germany, 2015; Volume 9351, pp. 234–241. [Google Scholar]
  108. Paszke, A.; Gross, S.; Massa, F.; Lerer, A.; Bradbury, J.; Chanan, G.; Killeen, T.; Lin, Z.; Gimelshein, N.; Antiga, L.; et al. PyTorch: An Imperative Style, High-Performance Deep Learning Library. In Proceedings of the Advances in Neural Information Processing Systems 2019 (NeurIPS); Curran Associates, Inc.: Red Hook, NY, USA, 2019; pp. 8024–8035. [Google Scholar]
  109. Yakubovskiy, P. Segmentation Models Pytorch. 2020. GitHub Repository. Available online: https://github.com/qubvel/segmentation_models.pytorch (accessed on 13 August 2021).
  110. Deng, J.; Dong, W.; Socher, R.; Li, L.J.; Li, K.; Fei-Fei, L. ImageNet: A large-scale hierarchical image database. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition 2009 (CVPR), Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar]
  111. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet Classification with Deep Convolutional Neural Networks. In Proceedings of the Advances in Neural Information Processing Systems 2012 (NIPS); Curran Associates, Inc.: Red Hook, NY, USA, 2012; pp. 1106–1114. [Google Scholar]
  112. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
  113. Tan, M.; Le, Q. EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks. Proc. Mach. Learn. Res. 2019, 97, 6105–6114. [Google Scholar]
  114. Roy, A.G.; Navab, N.; Wachinger, C. Recalibrating Fully Convolutional Networks With Spatial and Channel “Squeeze and Excitation” Blocks. IEEE Trans. Med. Imaging 2018, 38, 540–549. [Google Scholar] [CrossRef]
  115. Hu, J.; Shen, L.; Sun, G. Squeeze-and-Excitation Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition 2018 (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar]
  116. Roy, A.G.; Navab, N.; Wachinger, C. Concurrent Spatial and Channel ‘Squeeze & Excitation’ in Fully Convolutional Networks. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention 2018 (MICCAI); Lecture Notes in Computer Science; Springer International Publishing: Berlin/Heidelberg, Germany, 2018; pp. 421–429. [Google Scholar]
  117. Zagoruyko, S.; Komodakis, N. Wide Residual Networks. In Proceedings of the British Machine Vision Conference 2016 (BMVC); BMVA Press: New York, NY, USA, 2016; pp. 87.1–87.12. [Google Scholar]
  118. Tan, M.; Chen, B.; Pang, R.; Vasudevan, V.; Sandler, M.; Howard, A.; Le, Q.V. MnasNet: Platform-Aware Neural Architecture Search for Mobile. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition 2019 (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 2820–2828. [Google Scholar]
  119. Zhou, B.; Khosla, A.; Lapedriza, A.; Oliva, A.; Torralba, A. Learning Deep Features for Discriminative Localization. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition 2016 (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 2921–2929. [Google Scholar]
  120. Cohen, J. A coefficient of agreement for nominal scales. Educ. Psychol. Meas. 1960, 20, 37–46. [Google Scholar] [CrossRef]
  121. Chicco, D.; Jurman, G. The advantages of the Matthews Correlation Coefficient (MCC) over F1 score and accuracy in binary classification evaluation. BMC Genom. 2020, 21, 6. [Google Scholar] [CrossRef] [Green Version]
  122. Gorodkin, J. Comparing two K-category assignments by a K-category correlation coefficient. Comput. Biol. Chem. 2004, 28, 367–374. [Google Scholar] [CrossRef]
  123. Jurman, G.; Riccadonna, S.; Furlanello, C. A Comparison of MCC and CEN Error Measures in Multi-Class Prediction. PLoS ONE 2012, 7, e41882. [Google Scholar] [CrossRef] [Green Version]
  124. Wright, L. New Deep Learning Optimizer, Ranger: Synergistic Combination of RAdam + LookAhead for the Best of Both. 2020. Available online: https://medium.com/@lessw/new-deep-learning-optimizer-ranger-synergistic-combination-of-radam-lookahead-for-the-best-of-2dc83f79a48d (accessed on 13 August 2021).
  125. Wright, L. Ranger-Deep-Learning-Optimizer. 2020. Available online: https://github.com/lessw2020/Ranger-Deep-Learning-Optimizer (accessed on 13 August 2021).
  126. Zhang, M.; Lucas, J.; Ba, J.; Hinton, G.E. Lookahead Optimizer: k steps forward, 1 step back. In Proceedings of the Advances in Neural Information Processing Systems 2019 (NeurIPS); Curran Associates Inc.: Red Hook, NY, USA, 2019; pp. 9597–9608. [Google Scholar]
  127. Liu, L.; Jiang, H.; He, P.; Chen, W.; Liu, X.; Gao, J.; Han, J. On the Variance of the Adaptive Learning Rate and Beyond. In Proceedings of the International Conference on Learning Representation 2019 (ICLR), New Orleans, LA, USA, 6–9 May 2019; pp. 1–14. [Google Scholar]
  128. Kingma, D.P.; Ba, J. Adam: A Method for Stochastic Optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  129. Haibe-Kains, B.; Adam, G.A.; Hosny, A.; Khodakarami, F.; Waldron, L.; Wang, B.; McIntosh, C.; Goldenberg, A.; Kundaje, A.; Greene, C.S.; et al. Transparency and reproducibility in artificial intelligence. Nature 2020, 586, E14–E16. [Google Scholar] [CrossRef]
  130. The MAQC Consortium. The MAQC-II Project: A comprehensive study of common practices for the development and validation of microarray-based predictive models. Nat. Biotechnol. 2010, 28, 827–838. [Google Scholar] [CrossRef] [PubMed]
  131. Frank, S.J. Resource-frugal classification and analysis of pathology slides using image entropy. Biomed. Signal Process. Control 2021, 66, 102388. [Google Scholar] [CrossRef]
  132. Otsu, N. A Threshold Selection Method from Gray-Level Histograms. IEEE Trans. Syst. Man Cybern. 1979, 9, 62–66. [Google Scholar] [CrossRef] [Green Version]
  133. Romero-Zaliz, R.; Reinoso-Gordo, J.F. An Updated Review on Watershed Algorithms. In Soft Computing for Sustainability Science; Springer: Berlin/Heidelberg, Germany, 2017; pp. 235–258. [Google Scholar]
  134. Kuhn, H.W. The Hungarian Method for the assignment problem. Nav. Res. Logist. 1955, 2, 83–97. [Google Scholar] [CrossRef] [Green Version]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.