Next Article in Journal
Fast Phylogeny of SARS-CoV-2 by Compression
Next Article in Special Issue
Correction: Brochet et al. A Quantitative Comparison between Shannon and Tsallis–Havrda–Charvat Entropies Applied to Cancer Outcome Prediction. Entropy 2022, 24, 436
Previous Article in Journal
Distributed Quantization for Partially Cooperating Sensors Using the Information Bottleneck Method
Previous Article in Special Issue
Reconstructing Binary Signals from Local Histograms
 
 
Correction published on 13 May 2022, see Entropy 2022, 24(5), 685.
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Quantitative Comparison between Shannon and Tsallis–Havrda–Charvat Entropies Applied to Cancer Outcome Prediction

1
LITIS, Quantif, University of Rouen, 76000 Rouen, France
2
Centre Henri Becquerel, 76038 Rouen, France
3
Société Aquilab, 59120 Lille, France
4
Département de Radiothérapie, Centre Oscar Lambret, 59000 Lille, France
5
Département de Radiothérapie, CLCC Francois Baclesse, 14000 Caen, France
6
Département de Radiothérapie, Centre Léon Berard, 69008 Lyon, France
*
Author to whom correspondence should be addressed.
Entropy 2022, 24(4), 436; https://doi.org/10.3390/e24040436
Submission received: 7 February 2022 / Revised: 18 March 2022 / Accepted: 18 March 2022 / Published: 22 March 2022 / Corrected: 13 May 2022
(This article belongs to the Special Issue Application of Entropy to Computer Vision and Medical Imaging)

Abstract

:
In this paper, we propose to quantitatively compare loss functions based on parameterized Tsallis–Havrda–Charvat entropy and classical Shannon entropy for the training of a deep network in the case of small datasets which are usually encountered in medical applications. Shannon cross-entropy is widely used as a loss function for most neural networks applied to the segmentation, classification and detection of images. Shannon entropy is a particular case of Tsallis–Havrda–Charvat entropy. In this work, we compare these two entropies through a medical application for predicting recurrence in patients with head–neck and lung cancers after treatment. Based on both CT images and patient information, a multitask deep neural network is proposed to perform a recurrence prediction task using cross-entropy as a loss function and an image reconstruction task. Tsallis–Havrda–Charvat cross-entropy is a parameterized cross-entropy with the parameter α . Shannon entropy is a particular case of Tsallis–Havrda–Charvat entropy for α = 1 . The influence of this parameter on the final prediction results is studied. In this paper, the experiments are conducted on two datasets including in total 580 patients, of whom 434 suffered from head–neck cancers and 146 from lung cancers. The results show that Tsallis–Havrda–Charvat entropy can achieve better performance in terms of prediction accuracy with some values of α .

1. Introduction

This paper is devoted to studying the loss function based on Tsallis–Havrda–Charvat entropy in deep neural networks [1] for the prediction of outcomes in lung and head–neck cancers. When used for categorical prediction, the loss function is generally a cross-entropy related to a given entropy function. Indeed, cross-entropy-based loss functions are appropriate for evaluating how a probability distribution is close to the Dirac distribution. In deep neural networks, the output is a probability for each class obtained from a softmax activation function, and the Dirac distribution concentrated on one class represents the ground truth. There are several ways to compare these distributions [2,3,4]. Entropy-based metrics, such as divergences and cross-entropies, are the most common because they are the most appropriate way to sum up the informative content of a distribution, as explained in [5]. In Ref. [6], different entropies are presented. In classification and prediction, the cross-entropy is derived from an entropy measure and used as a loss function measuring the difference between the predicted probability and the real Dirac probability. To sum up, in most neural networks used for prediction, Shannon-related cross-entropy is the most common and widely used for segmentation [7], classification [8], or detection [9] and many other applications [10,11,12,13]. The reason why Shannon is the most used is twofold: first, because Shannon was the first entropy in the domain of information theory, and secondly, because it is extensive in the sense that the entropy of a multivariate distribution whose margins are independent is the sum of the marginal entropies. This last property makes the calculation of Shannon entropy easy. In Ref. [14], different ways of choosing an entropy and an associated divergence are detailed. Among them, Shannon entropy can be extended by replacing the logarithm by another function. Cross-entropies can be defined by replacing the counting measure (resp. Lebesgue measure for continuous case) by a Radon–Nykodim derivative between probability measures. Shannon’s entropy can be generalized on other entropies such as Renyi [15] and Tsallis–Havrda–Charvat [16,17]. In this paper, we are interested in a particular generalization of Shannon cross-entropy: Tsallis–Havrda–Charvat cross-entropy [18]. This class of entropies has the particularity of being parametrized with one parameter α and we recover Shannon entropy when the value of the parameter equal to 1. The relevance and possibilities of Tsallis–Havrda–Charvat in the medical field have been discussed before and this paper expands on the Tsallis–Havrda–Charvat formula.
Tsallis–Havrda–Charvat entropy was introduced independently by Tsallis [19] in the context of statistical physics and by Havrda and Charvat [20] in the context of information theory. Tsallis–Havrda–Charvat entropy has been used in publications in several fields, including medical imaging [21,22]. Tsallis–Havrda–Charvat entropy has rarely been used in deep learning, especially because of the difficulties with interpreting the hyperparameters α . However, there exist some scientific articles on this issue. In Ref. [23], Tsallis–Havrda–Charvat entropy is used to reduce the loss while classifying an image. In Ref. [18], the authors define Tsallis–Havrda–Charvat entropy in terms of axiomatization and propose a generalization based on this. In Ref. [24], the maximization of the entropy measure is studied for different classes of entropies such as Tsallis–Havrda–Charvat; the maximization of Tsallis–Havrda–Charvat entropy under constraints appears to be a way to generalize Gaussian distributions. In our previous work [17], Tsallis–Havrda–Charvat cross-entropy is used for the detection of noisy images in pulmonary microendoscopy. To capitalize and improve on this previous work, we propose to use deep learning in this paper. It allows for taking advantage of the previously used architecture by tuning and improving it to achieve better results. Deep learning is also very relevant for our field of study.
Deep learning has been widely developed in the medical field for classification or segmentation tasks [25,26,27]. Classification can be used to identify automatically the kind of cancer from which the patient is suffering [28,29] or the relevant outcomes after treatment, such as survival expectation [30] or relation to the treatment [31]. Recurrence in cancer after treatment is one of the main concerns for physicians [32], as it can dramatically impact the outcome for patients and their life expectancy. It would be beneficial for treatment selection if one could predict whether a recurrence will occur. Some studies have been carried out using CT scan images and clinical data. To our knowledge, there is no article using Tsallis–Havrda–Charvat for recurrence prediction [17]. The novelty of this article lies in the performance comparison of Shannon and Tsallis–Havrda–Charvat entropies in the context of cancer recurrence prediction with CT scan data combined with clinical information for patients affected by head and neck (H&N) or lung cancers (examples of the analyzed CT images are displayed in Figure 1). Moreover, we decided to study the parameter value in particular to examine its impact in order to predict these recurrences in both kinds of cancer. As medical data are generally scarce, the choice of a good entropy is important even if it can improve the performance only by 1 or 2 percent.
The paper is organized as follows. In the first section, we recall how categorical Shannon cross-entropy is defined and how it can be generalized for Tsallis–Havrda–Charvat. The second section is devoted to experiments and a comparison between the two entropies.

2. Entropy

As our problem concerns binary prediction, we focused only on finite-state random variables whose state-space was provided with the counting measure. Obviously, these results can be generalized for the finite-dimensional vectorial space R n provided with the Lebesgue measure.

2.1. Shannon Entropy and Related Cross-Entropy

For a discrete random variable Y, taking its values in Ω = { 1 , , k } with respective probabilities p 1 , , p k , Shannon entropy is defined by:
H ( p ) = i = 1 k log ( p i ) × p i
Shannon is minimal and equal to 0 if p i = 1 for one i and 0 otherwise; it is maximal if Y is uniformly distributed. The corresponding cross-entropy is given by:
H ( q : p ) = i = 1 k log ( q i ) p i
Generally, in a classification problem, the true distribution p is a Dirac distribution δ i 0 , where i 0 is the class of the data. In this case, the cross-entropy is H ( p ; q ) = log ( q i 0 ) 0 . As a consequence, q i 0 is closer to 0 and H ( p ; q ) is higher. Furthermore, minimizing the cross-entropy forces q to be as close as possible to the distribution p, and as this last one is the Dirac distribution δ i 0 , the minimum of the cross-entropy is 0.

2.2. Tsallis–Havrda–Charvat Cross-Entropy

There are several ways to generalize Shannon entropy, as explained in [14]. Shannon entropy can be expressed by:
H ( q ) = i = 1 k h ( q i )
where h ( u ) = u log ( u ) . h is a convex function such that h ( 1 ) = 0 . The idea is to choose another function satisfying the same properties. Tsallis–Havrda–Charvat entropy is defined by choosing:
h α ( u ) = u α u α 1 ,
where α > 0 and is given by:
H α ( q ) = 1 α 1 × 1 i = 1 k q i α
The associated cross-entropy is given by:
H α ( q : p ) = 1 α 1 × 1 i = 1 k q i α 1 p i
As for classical cross-entropy, Tsallis–Havrda–Charvat cross-entropy forces the predicted distribution q to be as close as possible to p when p is a Dirac distribution.

3. Neural Network Architecture for Recurrence Prediction

The proposed architecture used for recurrence prediction is a multitask neural network with a U-Net backbone, with one branch able to perform prediction tasks and another to reconstruct the input image for extracting features to help prediction. The architecture is presented in Figure 2.
The U-Net backbone is composed of three convolutional layers skipped by concatenations to add the features extracted from descending convolutions to the ascending ones. Within the convolution layers, we use ReLU activation functions.
At the bottom of the network, the extracted features are sent as inputs to one branch of fully connected layers in charge of making a decision path to determine whether the patient is at risk of recurrence.
Two main tasks are jointly carried out by the network. T1 is the reconstruction task, specific to the U-Net part of the network. It allows for determining whether the features extracted by the descending part of the U are relevant for prediction and classification and representative of the whole CT scan at the same time. The loss function used in this task is the mean squared error. It is defined as follows:
L rec = 1 N n = 1 N y n y ^ n 2 ,
where y n represents the true data values for the n-th patient and y ^ n is the estimated output from the network, with . being the Euclidean norm and N the number of patients.
The mean squared error computes the squared distance between the predicted output and the input image volumes. This function allows for comparing the predicted images and the true ones voxel-by-voxel and to train the network to recompose the images from the extracted features.
T2 is the prediction task. It is used to determine, from the same input, whether the current patient is at risk of encountering a recurrence of their cancer. It is constructed with fully connected layers and ends up in a binary classification. The prediction task’s loss function was the subject of our tests. We compared two entropies through this task.
The first, and most commonly used, was Shannon’s binary cross-entropy.
L pred , 1 = 1 N n = 1 N p n × log ( q n ) + ( 1 p n ) × log ( 1 q n ) ,
where p n is the true class, p n = 1 if recurrence, p n = 0 otherwise, and q n is the estimated probability of recurrence, with N being the number of patients.
The second entropy was the generalized formula, Tsallis–Havrda–Charvat binary cross-entropy.
L pred , α = 1 α 1 × 1 1 N n = 1 N ( q n α 1 p n + ( 1 q n ) α 1 ( 1 p n ) )
Binary cross-entropies are loss functions that are able to compare binary predictions with ground truths, which makes them relevant for our binary labels.
The total loss function of the network is the sum of the two losses. The choice of this total loss function was motivated by different experiments in which we used different weights for the individual loss functions. It appears that the sum with equal weights provided the best results.
L t o t a l = L rec + L pred , α
The prediction branch is the subject of interest for this article, with the reconstruction task being used to help the prediction in the feature extraction step.
Regarding the execution time and its change owing to the increasing or decreasing complexity of the network, we proposed to lock our network at a certain level of complexity and find an acceptable execution time of 5–7 h for a number of epochs of 100. This was considered to be a good compromise between obtaining a sufficient number of epochs to achieve meaningful results and not obtaining so many as to cause overfitting. We decided to use convolution layers in our U-Net backbone because of the limited computational power of available machines and thus the need to limit the number of parameters in the network. Allowing for a few hours of computations enabled the experiments to be run at night so that the results could be available and ready for analysis the next day.

4. Experiments

4.1. Datasets

The datasets were composed of 580 patients, among which 434 suffered from head–neck cancer and 146 from lung cancer. Both datasets were small in size. We chose to conduct experiments on the two subsets (head–neck and lung) separately. Indeed, the optimal value of α depends on the kind of data used. Moreover, we had already tested the total dataset of 580 and the results were poor. We therefore chose to show only the results for separate datasets.
CT images used in the neural network were resized with an image resolution of 128 × 128 × 64 voxels. The patient information used as input data in the neural network was of two kinds, namely quantitative and qualitative, as shown in Table 1 and Table 2.
Our experiments consisted of comparing the accuracy of Tsallis–Havrda–Charvat and Shannon for both datasets.

4.2. Evaluation Method

Since the study was conducted on small datasets (434 and 146 patients), a result validation strategy was required. We proposed to use the k-fold cross-validation relevant for small data validation. In our work, we used a five-fold cross-validation.
The procedure unfolds as follows:
  • Shuffle the dataset;
  • Split it randomly into 5 subsets;
  • For each subset:
    • Take the subset as a test dataset;
    • Take the other sets as training data;
    • Fit the model to the training data and evaluate it on the test dataset;
    • Retain the evaluation score;
  • Summarize the skill of the model from the samples of model evaluation scores.
Furthermore, accuracy was proposed as a metric for evaluation. It consisted, in this case, in comparing the values of the ground truth and the prediction and summing all these occurrences over the size of the dataset.

4.3. Results

The results achieved during the tests are displayed in this section. Reconstructed images are mainly used in order to show the relevance of extracted features for the prediction. Therefore, their performance is not very important here, because the objective is the prediction of recurrence. The original images and reconstructed ones are featured in Figure 3. We can see that the reconstructed images are similar to the input images, meaning that our network is able to recover input images.
The loss of information generates uncertainty in each image. A possible improvement would be to use a fuzzy image processor to improve the quality of the obtained images, as described in [33].

Comparison Results

Regarding Tsallis–Havrda–Charvat, we studied its hyperparameter α as varying from 0.1 to 2.0. When α = 1 , this entropy corresponds to Shannon entropy. The displayed p-value measures whether the results acquired by Tsallis–Havrda–Charvat five-fold cross-validation are statistically different from Shannon’s. Two conditions must be satisfied in order to accept the Tsallis–Havrda–Charvat entropy as providing better results than Shannon entropy: the average of the five-fold results must be superior to Shannon and the p-value must be smaller than 0.05 . The results are described in the following tables.
The results achieved for the dataset of head–neck cancers are described in Table 3.
Regarding the dataset containing lung cancers, the results achieved are described in Table 4.
After fine-tuning, we obtained a set of optimal hyperparameters.
The results achieved via the Tsallis–Havrda–Charvat formula confirm that, for most values of the hyperparameter α , the final accuracy is not superior to the accuracy obtained by Shannon’s loss function. It can also be observed that the loss function derived from Havrda–Charvat equation can provide better results than Shannon’s in some cases. However, it is difficult to know a priori what value of α is good for an application. Its choice is still a challenge.
We highlighted the results providing both better accuracy and a significant p-value in blue. The most promising values were achieved for α equal to 1.5 with head–neck cancers and between 1.9 and 3.5 with lung cancer. In this regard, we can state that the results obtained by Tsallis–Havrda–Charvat can be significantly improved compared to Shannon.
When analyzing the lung cancer results, a plateau can be easily noticed where Tsallis–Havrda–Charvat achieves better results than Shannon. This involves a specific set of values of α , from 1.9 to 3.5, in which the Tsallis–Havrda–Charvat loss function is significantly more efficient than Shannon’s.

5. Discussion

It has been determined that the Tsallis–Havrda–Charvat loss function performs equally well or better than Shannon cross-entropy, depending on the value of its hyperparameter α . It can be said that the Tsallis–Havrda–Charvat loss function, depending on the value of its hyperparameter α , can fit a wider array of input data and can potentially yield better results.
Conversely, we can state that, based on the calculated p-values and standard deviations, Tsallis–Havrda–Charvat entropy seems more unstable than Shannon entropy, as its standard deviation may reach 0.12, where Shannon’s is only 0.06. Furthermore, when looking at the p-values for several values of α , the results of Havrda–Charvat are not statistically different from Shannon’s. The instability of the results obtained using Tsallis–Havrda–Charvat could be explained by the fact that, despite the five-fold method, the data are still too scarce to reach a stable answer. In addition, data are tridimensional, making it more difficult than with 2D images to extract relevant features and thereby complicating the network’s tasks even further. In the analysis of 3D images, multiple slices must be taken into account in order to make a decision, which drastically increases the number of variables to be learned by the neural network. Moreover, in order to be usable for analysis by the neural network, the data are all supposed to have the same size. This is why, as for reasons of computational power, it was decided to have the data resized to 128 × 128 × 64 voxels. This size, despite still being full of information for the network, implies a loss of information for wider, larger and deeper images.
Nevertheless, the value of α plays a large part in the behavior of the loss function, and it is the key element that can be set to fit the input data, but the questions remains: what α fits which data? The choice of the value of the hyperparameter α remains a challenge as it depends strongly on the kind of data. In perspective, it would be interesting to develop an algorithm for selecting automatically the value of this hyperparameter in order to fit it as accurately as possible to the data. The aim is to reach the plateau, or area, of α where the Tsallis–Havrda–Charvat loss function provides consistently better results and features a smaller SD and p-value. A further analysis needs to be conducted on the link between input data and the location of the best α area in order to determine the kind of extracted feature and the kind of neuronal path that cause one area to be more efficient than another. For instance, in our case, the question is about which key feature of the input lung cancer images leads to better values between 1.9 and 3.5 and which key feature of the input H&N cancer images lead to better results for α equal to 1.5.

6. Conclusions

In this article, we established that, for our data and in some cases, Tsallis–Havrda–Charvat cross-entropy performs better than a Shannon-based loss function. Tsallis–Havrda–Charvat performed best on the head–neck dataset and lung dataset, at 80% and 81%, respectively, of correct recurrence prediction, while Shannon’s results for these two datasets were 67% and 52%, respectively. This makes the Tsallis–Havrda–Charvat formula the best candidate for further research on these datasets. For further research we might adapt Tsallis–Havrda–Charvat binary cross-entropy to a categorical cross-entropy. This would allow for making multi-class predictions, including estimating the time between the end of cancer treatment and recurrence. Another axis of evolution could be finding a way to automatically determine the proper value of α for a given application.

Author Contributions

Conceptualization, S.R. and J.L.-L.; methodology, S.R., T.B. and J.L.-L.; software, T.B. and J.L.-L.; validation, S.R.; formal analysis, T.B.; writing—original draft preparation, T.B.; writing—review and editing, J.L.-L. and S.R.; supervision, S.R.; project administration, S.R.; data curation, A.H., S.T., D.P., I.G., R.M., D.G., J.T., V.G. and P.V.; resources, S.T., D.P., I.G., R.M., J.T., V.G. and P.V.; medical expertise, P.V.; investigation, T.B., J.L.-L., A.H., S.T., D.P., I.G., R.M., D.G., J.T., V.G., P.V. and S.R. All authors have read and agreed to the published version of the manuscript.

Funding

This project was co-financed by the European Union with the European regional development fund (ERDF, 18P03390/18E01750/18P02733), by the Haute-Normandie Regional Council via the M2SINUM project and by the PRPO project from the Cancéropôle Nord-Ouest, France.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data are not publicly available due to being the property of Centre Henri Becquerel.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
MDPIMultidisciplinary Digital Publishing Institute
DOAJDirectory of open access journals
TLAThree-letter acronym
LDLinear dichroism
H&NHead–neck
H-CdHavrda–Charvat
GyGray

References

  1. Wang, Q.; Ma, Y.; Zhao, K.; Tian, Y. A Comprehensive Survey of Loss Functions in Machine Learning. Ann. Data Sci. 2020, 9, 187–212. [Google Scholar] [CrossRef]
  2. Chung, J.K.; Kannappan, C.T.; Sahoo, P.K. Measures of distance between probability distributions. J. Math. Anal. Appl. 1989, 138, 280–292. [Google Scholar] [CrossRef] [Green Version]
  3. Budzynski, R.J.; Kontracki, W.; Krolak, A. Applications of distance between distributions to gravitational wave data analysis. Class. Quantum Gravity 2007, 25, 015005. [Google Scholar] [CrossRef] [Green Version]
  4. Serrurier, M. An informational distance for estimating the faithfulness of a possibility distribution, viewed as a family of probability distributions, with respect to data. Int. J. Approx. Reason. 2013, 54, 919–933. [Google Scholar] [CrossRef]
  5. Kullback, S.; Leibler, R.A. On information and sufficiency. Ann. Math. Stat. 1951, 22, 79–86. [Google Scholar] [CrossRef]
  6. Amigó, J.M.; Balogh, S.G.; Hernández, S. A Brief Review of Generalized Entropies. Entropy 2018, 20, 813. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  7. Ma, Y.; Liu, Q.; Qian, Z.-B. Automated Image Segmentation Using Improved PCNN Model Based on Cross-entropy. In Proceedings of the 2004 International Symposium on Intelligent Multimedia, Video and Speech Processing, Hong Kong, China, 20–22 October 2004. [Google Scholar]
  8. Mannor, S.; Peleg, D.; Rubinstein, R. The cross entropy method for classification. In Proceedings of the 22nd International Conference on Machine Learning, Bonn, Germany, 7–11 August 2005. [Google Scholar]
  9. Qu, Z.; Mei, J.; Liu, L.; Zhou, D.Y. Crack Detection of Concrete Pavement With Cross-Entropy Loss Function and Improved VGG16 Network Model. In Proceedings of the 30th International Telecommunication Networks and Applications Conference (ITNAC), Melbourne, Australia, 25–27 November 2020. [Google Scholar]
  10. Silva, L.; Marques de Sá, J.; Alexandre, L.A. Neural network classification using Shannon’s entropy. In Proceedings of the ESANN 2005 Proceedings—13th European Symposium on Artificial Neural Networks, Bruges, Belgium, 27–29 April 2005; pp. 217–222. [Google Scholar]
  11. Rajinikanth, V.; Thanaraj, K.P.; Satapathy, S.C.; Fernandes, S.L.; Dey, N. Shannon’s Entropy and Watershed Algorithm Based Technique to Inspect Ischemic Stroke Wound. In Smart Intelligent Computing and Applications; Springer: Singapore, 2018; Volume 2, pp. 23–31. [Google Scholar]
  12. Ruby, U.; Yendapalli, V. Binary cross entropy with deep learning technique for Image classification. Int. J. Adv. Trends Comput. Sci. Eng. 2020, 9, 5393–5397. [Google Scholar] [CrossRef]
  13. Ramos, D.; Franco-Pedroso, J.; Lozano-Diez, A.; Gonzalez-Rodriguez, J. Deconstructing Cross-Entropy for Probabilistic Binary Classifiers. Entropy 2018, 20, 208. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  14. Basseville, M. Information: Entropies, Divergences et Moyenne; Technical Report; INRIA: Rennes, France, 2010. [Google Scholar]
  15. Andreieva, V.; Shvai, N. Generalization of Cross-Entropy Loss Function for Image Classification. Mohyla Math. J. 2021, 3, 3–10. [Google Scholar] [CrossRef]
  16. Roselin, R. Mammogram Image Classification: Non-Shannon Entropy based Ant-Miner. Int. J. Comput. Intell. Inf. 2014, 4, 33–43. [Google Scholar]
  17. Brochet, T.; Lapuyade-Lahorgue, J.; Bougleux, S.; Salaün, M.; Ruan, S. Deep Learning Using Havrda-Charvat Entropy for Classification of Pulmonary Optical Endomicroscopy. IRBM 2021, 42, 400–406. [Google Scholar] [CrossRef]
  18. Kumar, S.; Ram, G. A Generalization of the Havrda-Charvat and Tsallis Entropy and Its Axiomatic Characterization. Abstr. Appl. Anal. 2014, 2014, 505184. [Google Scholar] [CrossRef] [Green Version]
  19. Tsallis, C. Possible generalization of Boltzmann-Gibbs statistics. J. Stat. Phys. 1988, 52, 479–487. [Google Scholar] [CrossRef]
  20. Havrda, J.; Charvat, F. Quantification method of classification processes. Concept of structural alpha-entropy. Kybernetika 1967, 3, 30–35. [Google Scholar]
  21. Khader, M.; Ben Hamza, A. An Entropy-Based Technique for Nonrigid Medical Image Alignment. In Proceedings of the 14th International Conference on Combinatorial Image Analysis (IWCIA’11), Madrid, Spain, 23–25 May 2011; Springer: Berlin/Heidelberg, Germnay, 2011; pp. 444–455. [Google Scholar]
  22. Waleed, M.; Ben Hamza, A. Medical image registration using stochastic optimization. Opt. Lasers Eng. 2010, 48, 1213–1223. [Google Scholar] [CrossRef]
  23. Ramezani, Z.; Pourdarvish, A. Transfer learning using Tsallis entropy: An application to Gravity Spy. Phys. A Stat. Mech. Appl. 2021, 561, 125273. [Google Scholar] [CrossRef]
  24. Karmeshu, J. (Ed.) Entropy Measures, Maximum Entropy Principle and Emerging Applications; Springer: Berlin/Heidelberg, Germnay, 2003. [Google Scholar]
  25. Zhou, T.; Canu, S.; Vera, P.; Ruan, S. Latent Correlation Representation Learning for Brain Tumor Segmentation with Missing MRI Modalities. IEEE Trans. Image Process. 2021, 30, 4263–4274. [Google Scholar] [CrossRef]
  26. Amyar, A.; Modzelewski, R.; Li, H.; Ruan, S. Multi-task Deep Learning Based CT Imaging Analysis For COVID-19 Pneumonia: Classification and Segmentation. Comput. Biol. Med. 2020, 126, 104037. [Google Scholar] [CrossRef]
  27. Jin, L.; Shi, F.; Chun, Q.; Chen, H.; Ma, Y.; Wu, S.; Hameed, N.U.F.; Mei, C.; Lu, J.; Zhang, J.; et al. Artificial intelligence neuropathologist for glioma classification using deep learning on hematoxylin and eosin stained slide images and molecular markers. Neuro-Oncology 2021, 23, 44–52. [Google Scholar] [CrossRef]
  28. Sirinukunwattana, K.; Domingo, E.; Richman, S.D.; Redmond, K.L.; Blake, A.; Verrill, C.; Leedham, S.J.; Chatzipli, A.; Hardy, C.; Whalley, C.M.; et al. Image-based consensus molecular subtype (imCMS) classification of colorectal cancer using deep learning. Gut 2021, 70, 544–554. [Google Scholar] [CrossRef]
  29. Noorbakhsh, J.; Farahmand, S.; Foroughi, P.; Namburi, S.; Caruana, D.; Rimm, D.; Soltanieh-Ha, M.; Zarringhalam, K.; Chuang, J.H. Deep learning-based cross-classifications reveal conserved spatial behaviors within tumor histological images. Nat. Commun. 2020, 11, 6367. [Google Scholar] [CrossRef]
  30. Doppalapudi, S.; Qiu, R.G.; Badr, Y. Lung cancer survival period prediction and understanding: Deep learning approaches. Int. J. Med. Inform. 2021, 148, 104371. [Google Scholar] [CrossRef] [PubMed]
  31. Amyar, A.; Ruan, S.; Gardin, I.; Chatelain, C.; Decazes, P.; Modzelewski, R. 3D RPET-NET: Development of a 3D PET Imaging Convolutional Neural Network for Radiomics Analysis and Outcome Prediction. IEEE Trans. Radiat. Plasma Med. Sci. 2019, 3, 225–231. [Google Scholar] [CrossRef]
  32. Jiao, X.; Wang, Y.; Wang, F.; Wang, X. Recurrence pattern and its predictors for advanced gastric cancer after total gastrectomy. Medicine 2020, 99, e23795. [Google Scholar] [CrossRef] [PubMed]
  33. Versaci, M.; Calcagno, S.; Morabito, F.C. Image Contrast Enhancement by Distances Among Points in Fuzzy Hyper-Cubes. In Proceedings of the Part II 16th International Conference on Computer Analysis of Images and Patterns, CAIP 2015, Valletta, Malta, 2–4 September 2015; Springer: Berlin/Heidelberg, Germany, 2015; Volume 9257, pp. 494–505. [Google Scholar] [CrossRef]
Figure 1. Input images: head–neck CT (above) and lung CT (below).
Figure 1. Input images: head–neck CT (above) and lung CT (below).
Entropy 24 00436 g001
Figure 2. Architecture of the multitask neural network for recurrence prediction (T2) with the help of another task (T1: image reconstruction).
Figure 2. Architecture of the multitask neural network for recurrence prediction (T2) with the help of another task (T1: image reconstruction).
Entropy 24 00436 g002
Figure 3. Input images: original images (left) vs. reconstructed images (right).
Figure 3. Input images: original images (left) vs. reconstructed images (right).
Entropy 24 00436 g003
Table 1. Quantitative clinical data processed through the network.
Table 1. Quantitative clinical data processed through the network.
Clinical DataModality
Hemoglobing/dL
LymphocytesGiga/L
LeucocytesGiga/L
ThrombocytesGiga/L
Albuming/L
Treatment durationDays
Total irradiation doseGy
Number of fractions/
Average dose per fractionGy
Weight at the start and end of treatmentkg
Table 2. Qualitative clinical data processed through the network.
Table 2. Qualitative clinical data processed through the network.
Clinical DataModality
GenderM/F
TabacologySmoker, non-smoker, former smoker
Use of induction chemotherapyYes/No
Use of concomitant chemotherapyYes/No
TNMTumor, Node, Metastasis
Table 3. Accuracy obtained by loss function derived from Tsallis–Havrda–Charvat entropy in a function of α for the head–neck cancer dataset (p-values lower than 0.05 and accuracies higher than Shannon’s are highlighted in blue).
Table 3. Accuracy obtained by loss function derived from Tsallis–Havrda–Charvat entropy in a function of α for the head–neck cancer dataset (p-values lower than 0.05 and accuracies higher than Shannon’s are highlighted in blue).
5-Fold
α 12345AverageSDp-Value
0.10.680.530.60.580.630.600.060.01
0.30.600.700.700.700.430.630.120.28
0.50.580.580.600.700.730.640.070.27
0.70.850.700.600.700.650.700.090.25
0.90.580.600.600.600.680.610.040.07
1.00.750.650.680.580.700.670.06N/A (Shannon entropy)
1.10.680.750.750.730.750.730.030.09
1.30.630.730.680.750.700.700.050.32
1.50.700.780.850.800.880.800.070.03
1.70.800.730.630.750780.740.070.07
1.90.750.730.730.750.830.760.050.02
2.10.680.630.600.620.5750.630.040.12
2.30.730.730.730.730.70.720.010.09
2.50.750.580.680.680.60.660.070.34
2.70.680.530.450.630.60.580.090.04
2.90.730.730.730.750.730.730.010.07
3.10.70.550.650.570.630.620.060.02
3.30.650.650.650.580.550.620.050.07
3.50.730.750.730.750.700.730.020.08
3.70.730.700.600.600.580.640.070.20
3.90.680.700.550.580.530.610.080.09
Table 4. Accuracy obtained by loss function derived from Tsallis–Havrda–Charvat entropy in a function of α for the lung cancer dataset (p-values lower than 0.05 and accuracies higher than Shannon’s are highlighted in blue).
Table 4. Accuracy obtained by loss function derived from Tsallis–Havrda–Charvat entropy in a function of α for the lung cancer dataset (p-values lower than 0.05 and accuracies higher than Shannon’s are highlighted in blue).
5-Fold
α 12345AverageSDp-Value
0.10.580.580.470.580.630.570.060.23
0.30.580.580.580.580.680.600.040.13
0.50.630.580.520.580.520.560.030.26
0.70.580.580.630.630.580.600.030.13
0.90.630.530.680.470.530570.080.21
1.00.730.470.470.470.470.520.12N/A (Shannon entropy)
1.10.630.630.520.520.520.560.060.18
1.30.680.530.530.470.530.550.080.15
1.50.580.530.530.470.530.530.040.44
1.70.730.470.630.570.420.560.120.37
1.90.690.630.580.530.680.620.070.04
2.10.790.840.730.790.790.790.040.004
2.30.840.780.840.730.840.810.050.002
2.50.790.840.740.680.630.740.080.007
2.70.790.740.690.740.740.740.040.003
2.90.790.790.790.790.730.780.030.004
3.10.740.740.780.780.680.740.040.008
3.30.790.790.740.740.740.760.030.003
3.50.740.730.680.330.580.610.170.12
3.70.680.630.470.630.730.630.090.07
3.90.780.530.470.470.630.580.130.07
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Brochet, T.; Lapuyade-Lahorgue, J.; Huat, A.; Thureau, S.; Pasquier, D.; Gardin, I.; Modzelewski, R.; Gibon, D.; Thariat, J.; Grégoire, V.; et al. A Quantitative Comparison between Shannon and Tsallis–Havrda–Charvat Entropies Applied to Cancer Outcome Prediction. Entropy 2022, 24, 436. https://doi.org/10.3390/e24040436

AMA Style

Brochet T, Lapuyade-Lahorgue J, Huat A, Thureau S, Pasquier D, Gardin I, Modzelewski R, Gibon D, Thariat J, Grégoire V, et al. A Quantitative Comparison between Shannon and Tsallis–Havrda–Charvat Entropies Applied to Cancer Outcome Prediction. Entropy. 2022; 24(4):436. https://doi.org/10.3390/e24040436

Chicago/Turabian Style

Brochet, Thibaud, Jérôme Lapuyade-Lahorgue, Alexandre Huat, Sébastien Thureau, David Pasquier, Isabelle Gardin, Romain Modzelewski, David Gibon, Juliette Thariat, Vincent Grégoire, and et al. 2022. "A Quantitative Comparison between Shannon and Tsallis–Havrda–Charvat Entropies Applied to Cancer Outcome Prediction" Entropy 24, no. 4: 436. https://doi.org/10.3390/e24040436

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop