Next Article in Journal
Rolling Bearing Fault Diagnosis Based on WOA-VMD-MPE and MPSO-LSSVM
Next Article in Special Issue
The Role of Entropy in Construct Specification Equations (CSE) to Improve the Validity of Memory Tests: Extension to Word Lists
Previous Article in Journal
Progress in and Opportunities for Applying Information Theory to Computational Biology and Bioinformatics
Previous Article in Special Issue
MCMCINLA Estimation of Missing Data and Its Application to Public Health Development in China in the Post-Epidemic Era
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Unsupervised Hierarchical Classification Approach for Imprecise Data in the Breast Cancer Detection

by
Mario Fordellone
*,† and
Paolo Chiodini
Medical Statistics Unit, Universitiy of Campania “Luigi Vanvitelli”, 81100 Naples, Italy
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Entropy 2022, 24(7), 926; https://doi.org/10.3390/e24070926
Submission received: 23 May 2022 / Revised: 27 June 2022 / Accepted: 1 July 2022 / Published: 3 July 2022
(This article belongs to the Special Issue Statistical Methods for Medicine and Health Sciences)

Abstract

:
(1) Background: in recent years, a lot of the research of statistical methods focused on the classification problem in presence of imprecise data. A particular case of imprecise data is the interval-valued data. Following this research line, in this work a new hierarchical classification technique for multivariate interval-valued data is suggested for diagnosis of the breast cancer; (2) Methods: an unsupervised hierarchical classification method for imprecise multivariate data (called HC-ID) is performed for diagnosis of breast cancer (i.e., to discriminate between benign or malignant masses) and the results have been compared with the conventional (unsupervised) hierarchical classification approach (HC); (3) Results: the application on real data shows that the HC-ID procedure performs better HC procedure in terms of accuracy (HC-ID = 0.80, HC = 0.66) and sensitivity (HC-ID = 0.61, HC = 0.08). In the results obtained by the usual procedure, there is a high degree of false-negative (i.e., benign cancer diagnosis in malignant status) affected by the high degree of variability (i.e., uncertainty) characterizing the worst data.

1. Introduction

Among all types of cancers, breast cancer is one of the leading causes of death among middle-aged and old women. According to the World Health Organization (WHO), after two decades, lung cancer loses its sad record as the most widespread neoplasm. Breast cancer is now the most common oncological pathology. About 2.3 million new cases of breast cancer were diagnosed in 2020: 11.7% of all new cases of cancer [1,2].
Thus, prevention and an early diagnosis of breast tumors are immediate demands from society. Primary prevention is difficult as the causes of the disease are not well understood. However, if it can be detected at its early stage, the success rate of survival is quite high [3]. Physical examination and mammography are the best ways to make an early diagnosis of the disease. A precise detection, however, often depends on the visibility of microcalcifications in the mammogram. It is still challenging for radiologists to differentiate between benign and malignant cases. The existence of breast tumors is usually reflected in the mammogram. Some of the important signs of malignancy are: clustered calcifications, poorly defined masses, isolated dilated ducts, etc. However, all of these are not equally reflected in the mammograms [4].
Doctors physically look at the mammograms to detect deformations that may be taken as an indicator of cancerous changes and this could suffer from the human error and error with a visual inspection, which may further be enhanced by the poor quality of the mammogram images [5]. To try to solve these problems there is a demand for intelligent systems (e.g., statistical approaches, machine learning techniques, mathematical models, etc.) for early detection of tumors, assessment of their malignancy and monitoring of the same on the basis of multivariate features. In this direction, even some aiding tools would be of immense help. The efficiency and effectiveness of this process can be increased if tumors are detected and classified automatically through computers as benign or malignant [6].
In the breast cancer detection field, many classification approaches have been applied for diagnostic purposes, e.g., [6] use rank nearest neighbor (RNN) classification rules, in [7] authors focus on association rules (AR) and artificial neural network (ANN), in [8] authors propose a non-parametric statistical model, in [9] is shown a comparative study of machine learning algorithms, in [10] a novel approach using data mining techniques is presented, and in [11] authors use machine learning approaches as well as Naive Bayes (NB) classifier and k-nearest neighbor (KNN). For a detailed review, the reader can refer to [12]. Notice that all these proposals are based on supervised techniques. In the supervised learning model, the algorithms learn from labeled data (i.e., the structure groups are known). After understanding the data, the algorithm determines which label should be given to new data by associating patterns to the unlabeled new data [13]. In medical and statistical literature, little attention is paid to research works where unsupervised techniques for cancer detection have been proposed. In unsupervised learning, the algorithms segregate the data in a data set in which the data is unlabeled (i.e., the structure groups are unknown) based on some hidden features in the data. This function can be useful for discovering the hidden structure of data and for tasks like anomaly detection [14]. In [15], authors tried to predict the disease based on relevant features in the data through the use of unsupervised algorithms; in [16] authors used the K-means (KM) algorithm to evaluate the impact of clustering for the early detection of breast cancer, using centroid initialization, distance measures, and split methods; in [17] authors proposed a review based on several ultrasound image segmentation techniques, mainly focus on eight clustering methods over the last 10 years, and they showed the advantages and disadvantages of these approaches; in [18] authors proposed a comparative study where three different unsupervised learning models have been used for breast cancer detection.
However, in real-life applications, the results of measurements are never precise (i.e., some degree of uncertainty that characterizes them there is). The uncertainty of a measurement can be defined as the interval on the measurement scale within which the true value lies with a specified probability when all sources of error have been taken into account [19,20]. From a statistical point of view, in recent years the research of statistical methodologies to analyze complex structures of data has increased. In particular, a lot of attention has been focused on the imprecise data [21]. For example, the concentration of toxic substances in different environmental media are imprecise quantities and then, their measurements are not precise. In this work, an unsupervised hierarchical classification method for imprecise multivariate data (called HC-ID) is performed for the diagnosis of breast cancer (i.e., to discriminate between benign or malignant masses) and the results have been compared with the conventional (unsupervised) hierarchical classification approach (HC). Notice that both HC approaches are performed by the complete linkage [22] method and then, in agglomerative way. For other examples of HC application in breast cancer detection, the reader can refer to [23,24,25]. In recent years, the research of statistical methods to analyze complex structures of data has increased. In particular, a lot of attention has been focused on the unsupervised and supervised classification problem in presence of imprecise data [26,27,28,29]. The simplest case of imprecise data is the interval-valued data [26,30]. In the literature on data analysis, a great deal of attention is paid to statistical methods to treat interval-valued data, in different research areas [26,30,31,32,33,34]. The novelty of this work is to consider the variability (i.e., the uncertainty) of the data in the classification procedure. In many medical research areas, such as in cancer detection studies, the results can be affected by measurement uncertainty, and this, in turn, could affect the statistical analysis reliability. In these cases, researchers should be interested to consider the uncertainty as a crucial part of the information rather than a simple noise factor.

2. Materials and Methods

2.1. Methodology of the Proposed Approach

We can formalize an interval-valued data as x i j = [ x ̲ i j , x ¯ i j ] , i = 1 , , n and j = 1 , , J ; where x i j is the j-th interval valued variable observed on the i-th observation, x ̲ i j and x ¯ i j denote the lower and upper bounds of the interval, respectively, (i.e., the minimum and maximum values registered for the j-th interval-valued variable with respect to the i-th observation). Then, in an n × J interval-valued data matrix, each observation is represented as a hyperrectangle (in R J ) having 2 J vertices. However, a simpler notation of interval-valued data consists to consider centers and radii, separately. In particular, we can indicate C the n × J centers matrix (or midpoints matrix) whose generic element c i j = 2 1 ( x ̲ i j + x ¯ i j ) is the center (midpoint) of the associated interval. Furthermore, we can define R the n × J radii matrix whose generic element r i j = 2 1 ( x ¯ i j x ̲ i j ) is the radius of the associated interval. Then, by considering this reformulation of the interval-valued data, the interval-valued matrix can be formalized as follows:
X x i j = [ c i j , r i j ] : i = 1 , , n ; j = 1 , , J
In the left plot of Figure 1 is represented a bi-dimensional dataset in ordinary form (i.e., with a radius equal to zero), while in the right one is represented a bi-dimensional interval-valued dataset.
We can note that a structure of three groups characterizes our datasets. In particular, 300 observations by three different bi-variate normal distributions (i.e., 100 for each group) have been generated in order to obtain the left plots. Subsequently, other three bi-variate normal distributions have been used to obtain random radii for the right plot.
The generic interval-valued data pertaining to the i-th observation with respect to the j-th interval-valued feature can be shown as the pair ( c i j , r i j ), i = 1 , , n and j = 1 , , J , where c i j denotes the center and r i j the radius of the interval (i.e., x i j = c i j ± r i j ). In the literature, several metrics have been suggested for interval-valued.
Let I ¯ i j be the i-th interval with respect to the j-th feature, within an interval pair { I ¯ i j , I ¯ i j } , OR between I ¯ i j and I ¯ i j is defined as
O R ( I ¯ i j , I ¯ i j ) = | I ¯ i j I ¯ i j | | I ¯ i j | ,
where | I ¯ i j I ¯ i j | is the size of the intersection between I ¯ i j and I ¯ i j , while | I ¯ i j | is the size of interval I ¯ i j . OR for an interval in a given pair will fall under one of the following cases:
  • O R ( I ¯ i j , I ¯ i j ) = 1 , when I ¯ i j and I ¯ i j are identical;
  • O R ( I ¯ i j , I ¯ i j ) = 0 , when I ¯ i j and I ¯ i j are disjointed;
  • Otherwise, 0 O R ( I ¯ i j , I ¯ i j ) 1 .
Thus, the overlapping ratio-based similarity measure S O R takes into consideration the reciprocal similarity of intervals within a pair in order to estimate their overall similarity. Formally, S O R for a pair of intervals, I ¯ i j and I ¯ i j , is the vectors sum (i.e., the n o r m ) of their overlapping ratios:
S O R ( I ¯ i j , I ¯ i j ) = O R ( I ¯ i j , I ¯ i j ) , O R ( I ¯ i j , I ¯ i j ) = | I ¯ i j I ¯ i j | | I ¯ i j | , | I ¯ i j I ¯ i j | | I ¯ i j | .
Then, the overlapping ratio-based similarity measure for a pair of intervals I ¯ i and I ¯ i in the J-dimensional space, is defined as
S O R J ( I ¯ i , I ¯ i ) = j = 1 J | I ¯ i j I ¯ i j | | I ¯ i j | + | I ¯ i j I ¯ i j | | I ¯ i j | 2 J .
Note that in this work a distance measure D O R J ( I ¯ i , I ¯ i ) [35] has been used, which can easily be derived as
D O R J ( I ¯ i , I ¯ i ) = j = 1 J 1 | I ¯ i j I ¯ i j | | I ¯ i j | + 1 | I ¯ i j I ¯ i j | | I ¯ i j | 2 J .
The final result is a distance matrix characterized by intervals. In this work, we use this particular distance matrix for the classification purpose of imprecise data. In particular, a hierarchical classification method with an interval-valued distance matrix and a complete linkage approach has been performed. This new model is called HC-ID.

2.2. Description of the Breast Cancer Example Data

In this work, an analysis of the Breast Cancer Wisconsin (Diagnostic) dataset is performed (https://archive.ics.uci.edu/ml/datasets/breast+cancer+wisconsin+(diagnostic) accessed on 27 June 2022). This data set was created by [36] and it has been very used for training statistical methods (e.g., [37]). To create the dataset Dr. Wolberg used fluid samples, taken from patients with solid breast masses and an easy-to-use graphical computer program called Xcyt, which is capable of performing the analysis of cytological features based on a digital scan. The program uses a curve-fitting algorithm, to compute ten features from each one of the cells in the sample, then it calculates the mean value, extreme value and standard error of each feature for the image, returning a 30 real-valued vector. Dataset consists of 569 patients, 357 with benign diagnosis and 212 with malignant status.
Attribute Information (response variable):
  • Diagnosis (M = malignant, B = benign).
Ten real-valued features are computed for each cell nucleus:
1.
Radius (mean of distances from center to points on the perimeter);
2.
Texture (standard deviation of gray-scale values);
3.
Perimeter;
4.
Area;
5.
Smoothness (local variation in radius lengths);
6.
Compactness (perimeter 2 /area − 1.0);
7.
Concavity (severity of concave portions of the contour);
8.
Concave points (number of concave portions of the contour);
9.
Symmetry;
10.
Fractal dimension (“coastline approximation” − 1).
The mean, standard error and “worst” or largest (mean of the three largest values) of these features were computed for each image, resulting in 30 features. For instance, field 3 is the mean radius, field 13 is radius se, and field 23 is the worst radius. For breast cancer diagnosis we have compared the results obtained by the proposed HC-ID approach with those one obtained by the conventional HC approach. In particular, to obtain the HC-ID model the dissimilarity measure for interval-valued data based on the overlapping ratio (OR) proposed by [35] has been applied to the interval-valued dataset. Then, we have used the 10 worst features (i.e., the feature with greater uncertainty/variability) as the centers of the interval data, while the standard deviations (i.e., the degree of uncertainty) are the radii. In this way, the classification procedure includes also the degree of uncertainty (i.e., variability) characterizing data and more homogeneous and separated groups are guaranteed (for details on the imprecise data concept, the reader can refer to [26,28,38]). Notice that for the classification we assume that the observed diagnosis groups is unknown (i.e., unsupervised classification).

2.3. Statistical Analysis

For the hierarchical classification model the hclust R package was used, while to obtain the interval-valued distance matrix based on D O R J ( I ¯ i , I ¯ i ) , the reader can refer to web page: https://github.com/mfordellone/Unsupervised-hierarchical-classification-approach-for-imprecise-data-in-the-breast-cancer-detection.git accessed on 27 June 2022. Notice that HC-ID is an unsupervised technique and then, it helps the analyst to identify data-driven patterns that may warrant further investigation but the prediction is not provided. You can easily use HC-ID to perform clustering, and from there for every new data point, you just find which cluster it matches most closely.
To evaluate the diagnostic performance of the HC-ID model sensitivity, specificity, positive predictive value, negative predictive value, positive likelihood ratio (LR+), negative likelihood ratio (LR-), and the accuracy rate have been used.

3. Results

The dataset consists of 569 patients, where 212 (37.26%) have malignant breast cancer and 357 (62.74%) have benign. Figure 2 shows all the variables distribution (worst) included in the analysis with respect to the observed diagnosis groups of data (M: Malignant, B: Benign).
By applying HC-ID (i.e., the use of the interval-valued distance matrix obtained via OR approach) the automatic classification shown by the dendrogram in Figure 3 is obtained. The predicted diagnosis group proportions are 28.82% for malignant breast cancer and 71.18% for benign breast cancer. By applying the conventional HC (i.e., the use of data points to obtain the distance matrix of the worst data), the automatic classification shown by dendrogram in Figure 4 is obtained. In this case, the predicted diagnosis group proportions are 2.98% for malignant breast cancer and 97.02% for benign breast cancer.
The results show that the classification procedure based on interval-valued data performs better than the usual procedure in terms of accuracy and sensitivity. In the results obtained by the usual procedure, there is a high degree of false-negative (i.e., benign cancer diagnosis in malignant status) affected by the high degree of variability (i.e., uncertainty) characterizing the worst data. Additionally, the dendrograms represented in Figure 3 and Figure 4 show a greater homogeneity in the partition obtained by the interval-valued approach. Finally, Table 1 shows the summary of the performance obtained by the two approaches.
In columns 1 and 4 of the table are shown the estimated values of diagnostic evaluation measures obtained by HC-ID and HC approaches, respectively; in columns 2–3 and 5–6 are shown the estimated confidence intervals at 95% (i.e., Lower 95% and Upper 95% are the lower limits and the upper limit of the confidence interval) on the diagnostic evaluation measures obtained by HC-ID and HC approaches, respectively.

4. Discussion

In this work, an unsupervised hierarchical classification method for interval-valued multivariate data (HC-ID) is performed for diagnosis of the breast cancer. In particular, a methodology able to discriminate between benign or malignant breast masses has been proposed. Moreover, in order to show the good performance of the proposed classification model comparison with the conventional (unsupervised) hierarchical classification approach is carried out.
The principal novelty of the proposed approach is the use of an unsupervised classification approach. In fact, the most important previous proposals ([9,10,11], etc.) are based on the use of classification methodologies where the observed diagnosis groups are known. However, in some real cases, this information could not be available.
For application purpose, an analysis of the Breast Cancer Wisconsin (Diagnostic) dataset (https://archive.ics.uci.edu/ml/datasets/breast+cancer+wisconsin+(diagnostic) accessed on 27 June 2022) created by [36] is performed. In particular, we have used the 10 worst features as the centers of the interval data, while the standard deviations are the radii. In this way, the classification procedure includes also the degree of uncertainty (i.e., variability) characterizing data and more homogeneous and separated groups are guaranteed. In fact, the results show that our proposal performs better than the conventional procedure (i.e., the HC approach) in terms of accuracy, sensitivity and negative predictive value. However, the specificity and the positive predictive value obtained by the usual procedure are equal to 1 but, unfortunately, also the false-negative rate increases. Moreover, the LR+ obtained by HC is characterized by high variability and shows a very large confidence interval. We think that this result is affected by the high degree of variability (i.e., uncertainty) characterizing the worst data. Moreover, we think that the high rate of false-negative in cancer detection fields is a serious problem. In particular, false-negative tests at diagnosis of early disease and of relapse resulted in diagnostic and therapeutic delays.
We think that the principal advantages of the HC-ID approach consist of (i) to include the uncertainty of the data in the classification procedure that leads to more homogeneous partitions of subjects; (ii) the possibility to consider a multi-group approach that encourages the use of the procedure for different purposes (e.g., stages detection or identification of prognosis classes); (iii) the external procedure of uncertainty estimation that leads to fix a different kind of measures (e.g., IQR, specific percentile differences, other intervals symmetrical or not symmetrical with respect the point data, etc.).
Whereas, the principal disadvantages consist of (i) the correct estimation of uncertainty since it is not simple; (ii) to fix a constant uncertainty measure of the subjects is a very strong assumption. The subjects could have some characteristics to affect the variability degree in different measures; (iii) the approach is not very adequate in cases with small sample sizes. In these cases, the radii of the imprecise datum could be very high and the risk to associate the biggest weight to the uncertainty than the point data is hard to handle. However, in Appendix A a validation study is proposed in order to study the HC-ID model behavior for different sample size.
Finally, we think that our proposed approach is very useful for cancer diagnostic purposes in the cases where there is a marked variability in the subjects’ features and where the outcome information is incomplete or not available. For future research, could be interesting to provide a validation of the results using other datasets and other cancer types, because at moment this is a real limit of this work. Additionally, the comparison with other classification statistical models could be an interesting development for this research line.

Author Contributions

Conceptualization, M.F.; methodology, M.F. and P.C.; software, M.F.; validation, P.C.; formal analysis, M.F. and P.C.; writing—original draft preparation, P.C.; writing—review and editing, P.C.; visualization, M.F.; supervision, P.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
MDPIMultidisciplinary Digital Publishing Institute
DOAJDirectory of open access journals
TLAThree letter acronym
LDLinear dichroism

Appendix A. Validation Study

In order to study the HC-ID model behavior for different sample size, the model has been applied on two sub-samples of 150 and 300 subjects random selected, without replacement, from the Breast Cancer Wisconsin dataset. Then, in the first case dataset consists of 150 patients, where 55 (36.67%) have a malignant breast cancer and 95 (63.33%) have the one benign; in the second case dataset consists of 300 patients, where 114 (37.26%) have a malignant breast cancer and 186 (62.74%) have the one benign. Table A1 and Table A2 show the results obtained by HC-ID model compared with those one obtained by the conventional HC model.
Table A1. Comparison of the performances obtained by interval-valued approach and conventional approach on a randomized sub-sample of 150 subjects.
Table A1. Comparison of the performances obtained by interval-valued approach and conventional approach on a randomized sub-sample of 150 subjects.
Interval-Valued ApproachConventional Approach
EstimateLower 95%Upper 95%EstimateLower 95%Upper 95%
Sensitivity1.0000.9621.0000.0360.0040.125
Specificity0.7090.5710.8241.0000.9621.000
Pos.Pred.Val.0.8560.7760.9151.0000.1581.000
Neg.Pred.Val.1.0000.9101.0000.6420.5590.719
LR+3.4382.2755.1938.5710.419175.363
LR−0.0070.0000.1180.9640.9151.014
Accuracy0.8930.8330.9380.6470.5650.723
Table A2. Comparison of the performances obtained by interval-valued approach and conventional approach on a randomized sub-sample of 300 subjects.
Table A2. Comparison of the performances obtained by interval-valued approach and conventional approach on a randomized sub-sample of 300 subjects.
Interval-Valued ApproachConventional Approach
EstimateLower 95%Upper 95%EstimateLower 95%Upper 95%
Sensitivity0.9350.8900.9660.0090.0000.048
Specificity0.6230.5270.7121.0000.9801.000
Pos.Pred.Val.0.8020.7430.8531.0000.0251.000
Neg.Pred.Val.0.8550.7610.9230.6220.5640.677
LR+2.4801.9533.1494.8780.200118.743
LR−0.1040.0590.1820.9910.9741.008
Accuracy0.8170.7680.8590.6230.5660.678
The two tables show that HC-ID performs better than the conventional HC model except in specificity and positive predictive value. In particular, we can see that the results are not affected by the sample size since they are very similar than the results reported in Table 1.

References

  1. Tao, Z.; Shi, A.; Lu, C.; Song, T.; Zhang, Z.; Zhao, J. Breast cancer: Epidemiology and etiology. Cell Biochem. Biophys. 2015, 72, 333–338. [Google Scholar] [CrossRef] [PubMed]
  2. Huang, J.; Chan, P.S.; Lok, V.; Chen, X.; Ding, H.; Jin, Y.; Yuan, J.; Lao, X.Q.; Zheng, Z.J.; Wong, M.C. Global incidence and mortality of breast cancer: A trend analysis. Aging 2021, 13, 5748. [Google Scholar] [CrossRef] [PubMed]
  3. Wang, L. Early diagnosis of breast cancer. Sensors 2017, 17, 1572. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Al-Dhabyani, W.; Gomaa, M.; Khaled, H.; Fahmy, A. Dataset of breast ultrasound images. Data Brief 2020, 28, 104863. [Google Scholar] [CrossRef]
  5. Kolb, T.M.; Lichy, J.; Newhouse, J.H. Comparison of the performance of screening mammography, physical examination, and breast US and evaluation of factors that influence them: An analysis of 27,825 patient evaluations. Radiology 2002, 225, 165–175. [Google Scholar] [CrossRef] [Green Version]
  6. Bagui, S.C.; Bagui, S.; Pal, K.; Pal, N.R. Breast cancer detection using rank nearest neighbor classification rules. Pattern Recognit. 2003, 36, 25–34. [Google Scholar] [CrossRef]
  7. Karabatak, M.; Ince, M.C. An expert system for detection of breast cancer based on association rules and neural network. Expert Syst. Appl. 2009, 36, 3465–3469. [Google Scholar] [CrossRef]
  8. Cheng, H.D.; Shan, J.; Ju, W.; Guo, Y.; Zhang, L. Automated breast cancer detection and classification using ultrasound images: A survey. Pattern Recognit. 2010, 43, 299–317. [Google Scholar] [CrossRef] [Green Version]
  9. Bazazeh, D.; Shubair, R. Comparative study of machine learning algorithms for breast cancer detection and diagnosis. In Proceedings of the 2016 5th International Conference on Electronic Devices, Systems and Applications (ICEDSA), Ras Al Khaimah, United Arab Emirates, 6–8 December 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 1–4. [Google Scholar]
  10. Chaurasia, V.; Pal, S. A novel approach for breast cancer detection using data mining techniques. Int. J. Innov. Res. Comput. Commun. Eng. 2017, 2, 1–17. [Google Scholar]
  11. Amrane, M.; Oukid, S.; Gagaoua, I.; Ensari, T. Breast cancer classification using machine learning. In Proceedings of the 2018 Electric Electronics, Computer Science, Biomedical Engineerings’ Meeting (EBBT), Istanbul, Turkey, 18–19 April 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 1–4. [Google Scholar]
  12. Ramadan, S.Z. Methods used in computer-aided diagnosis for breast cancer detection using mammograms: A review. J. Healthc. Eng. 2020, 2020, 9162464. [Google Scholar] [CrossRef] [Green Version]
  13. Carrizosa, E.; Morales, D.R. Supervised classification and mathematical optimization. Comput. Oper. Res. 2013, 40, 150–165. [Google Scholar] [CrossRef] [Green Version]
  14. Bandyopadhyay, S.; Saha, S. Unsupervised Classification: Similarity Measures, Classical and Metaheuristic Approaches, and Applications; Springer: Berlin/Heidelberg, Germany, 2013. [Google Scholar]
  15. Gharibdousti, M.S.; Haider, S.M.; Ouedraogo, D.; Susan, L. Breast cancer diagnosis using feature extraction techniques with supervised and unsupervised classification algorithms. Appl. Med. Inform. 2019, 41, 40–52. [Google Scholar]
  16. Dubey, A.K.; Gupta, U.; Jain, S. Analysis of k-means clustering approach on the breast cancer Wisconsin dataset. Int. J. Comput. Assist. Radiol. Surg. 2016, 11, 2033–2047. [Google Scholar] [CrossRef] [PubMed]
  17. Muhammad, M.; Zeebaree, D.; Brifcani, A.M.A.; Saeed, J.; Zebari, D.A. Region of interest segmentation based on clustering techniques for breast cancer ultrasound images: A review. J. Appl. Sci. Technol. Trends 2020, 1, 78–91. [Google Scholar]
  18. Niţică, Ş.; Czibula, G.; Tomescu, V.I. A comparative study on using unsupervised learning based data analysis techniques for breast cancer detection. In Proceedings of the 2020 IEEE 14th International Symposium on Applied Computational Intelligence and Informatics (SACI), Timisoara, Romania, 21–23 May 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 000099–000104. [Google Scholar]
  19. Analytical Methods Committee. Uncertainty of measurement: Implications of its use in analytical science. Analyst 1995, 120, 2303–2308. [Google Scholar] [CrossRef]
  20. Oosterhuis, W.P.; Bayat, H.; Armbruster, D.; Coskun, A.; Freeman, K.P.; Kallner, A.; Koch, D.; Mackenzie, F.; Migliarino, G.; Orth, M.; et al. The use of error and uncertainty methods in the medical laboratory. Clin. Chem. Lab. Med. CCLM 2018, 56, 209–219. [Google Scholar] [CrossRef]
  21. Bandemer, H.; Näther, W. Fuzzy Data Analysis; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2012; Volume 20. [Google Scholar]
  22. Hartigan, J.A. Statistical theory in clustering. J. Classif. 1985, 2, 63–76. [Google Scholar] [CrossRef]
  23. Makretsov, N.A.; Huntsman, D.G.; Nielsen, T.O.; Yorida, E.; Peacock, M.; Cheang, M.C.; Dunn, S.E.; Hayes, M.; van de Rijn, M.; Bajdik, C.; et al. Hierarchical clustering analysis of tissue microarray immunostaining data identifies prognostically significant groups of breast carcinoma. Clin. Cancer Res. 2004, 10, 6143–6151. [Google Scholar] [CrossRef] [Green Version]
  24. Triayudi, A.; Fitri, I. Comparison of parameter-free agglomerative hierarchical clustering methods. ICIC Express Lett. 2018, 12, 973–980. [Google Scholar]
  25. Huang, Z.; Chen, D. A Breast Cancer Diagnosis Method based on VIM Feature Selection and Hierarchical Clustering Random Forest Algorithm. IEEE Access 2021, 10, 3284–3293. [Google Scholar] [CrossRef]
  26. D’Urso, P.; De Giovanni, L. Robust clustering of imprecise data. Chemom. Intell. Lab. Syst. 2014, 136, 58–80. [Google Scholar] [CrossRef]
  27. Coppi, R.; D’Urso, P.; Giordani, P. Fuzzy and possibilistic clustering for fuzzy data. Comput. Stat. Data Anal. 2012, 56, 915–927. [Google Scholar] [CrossRef]
  28. Hathaway, R.J.; Bezdek, J.C.; Pedrycz, W. A parametric model for fusing heterogeneous fuzzy data. IEEE Trans. Fuzzy Syst. 1996, 4, 270–281. [Google Scholar] [CrossRef]
  29. Sato, M.; Sato, Y. Fuzzy clustering model for fuzzy data. In Proceedings of the 1995 IEEE International Conference on Fuzzy Systems, Yokohama, Japan, 20–24 March 1995; IEEE: Piscataway, NJ, USA, 1995; Volume 4, pp. 2123–2128. [Google Scholar]
  30. D’Urso, P.; Leski, J. Fuzzy c-ordered medoids clustering for interval-valued data. Pattern Recognit. 2016, 58, 49–67. [Google Scholar] [CrossRef]
  31. Coppi, R.; Giordani, P.; D’Urso, P. Component models for fuzzy data. Psychometrika 2006, 71, 733. [Google Scholar] [CrossRef]
  32. D’Urso, P.; Giordani, P. A possibilistic approach to latent component analysis for symmetric fuzzy data. Fuzzy Sets Syst. 2005, 150, 285–305. [Google Scholar] [CrossRef]
  33. Giordani, P.; Kiers, H.A. Principal component analysis of symmetric fuzzy data. Comput. Stat. Data Anal. 2004, 45, 519–548. [Google Scholar] [CrossRef]
  34. Denoeux, T.; Masson, M. Multidimensional scaling of interval-valued dissimilarity data. Pattern Recognit. Lett. 2000, 21, 83–92. [Google Scholar] [CrossRef]
  35. Kabir, S.; Wagner, C.; Havens, T.C.; Anderson, D.T.; Aickelin, U. Novel similarity measure for interval-valued data based on overlapping ratio. In Proceedings of the 2017 IEEE International Conference on Fuzzy Systems (FUZZ-IEEE), Naples, Italy, 9–12 July 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 1–6. [Google Scholar]
  36. Mangasarian, O.L.; Street, W.N.; Wolberg, W.H. Breast cancer diagnosis and prognosis via linear programming. Oper. Res. 1995, 43, 570–577. [Google Scholar] [CrossRef] [Green Version]
  37. Agarap, A.F.M. On breast cancer detection: An application of machine learning algorithms on the wisconsin diagnostic dataset. In Proceedings of the 2nd International Conference on Machine Learning and Soft Computing, Phu Quoc Island, Vietnam, 2–4 February 2018; pp. 5–9. [Google Scholar]
  38. D’Urso, P.; Giordani, P. A weighted fuzzy c-means clustering model for fuzzy data. Comput. Stat. Data Anal. 2006, 50, 1496–1523. [Google Scholar] [CrossRef]
Figure 1. Artificial data generated by three bi-variate Normal distributions. To the left we have dataset in ordinary form; to the right we have interval-valued dataset.
Figure 1. Artificial data generated by three bi-variate Normal distributions. To the left we have dataset in ordinary form; to the right we have interval-valued dataset.
Entropy 24 00926 g001
Figure 2. Variables distribution with respect to the observed diagnosis groups of data.
Figure 2. Variables distribution with respect to the observed diagnosis groups of data.
Entropy 24 00926 g002
Figure 3. Dendrogram obtained by HC applied on interval-valued data.
Figure 3. Dendrogram obtained by HC applied on interval-valued data.
Entropy 24 00926 g003
Figure 4. Dendrogram obtained by HC applied on worst data.
Figure 4. Dendrogram obtained by HC applied on worst data.
Entropy 24 00926 g004
Table 1. Comparison of the performances obtained by interval-valued approach and conventional approach.
Table 1. Comparison of the performances obtained by interval-valued approach and conventional approach.
Interval-Valued ApproachConventional Approach
EstimateLower 95%Upper 95%EstimateLower 95%Upper 95%
Sensitivity0.6130.5440.6790.0800.0470.125
Specificity0.9050.8690.9331.0000.9901.000
Pos.Pred.Val.0.7930.7230.8521.0000.8051.000
Neg.Pred.Val.0.7980.7550.8360.6470.6050.687
LR+6.4394.5969.02058.8263.556973.204
LR−0.4270.3600.5080.9200.8830.957
Accuracy0.7960.7610.8290.6570.6170.696
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Fordellone, M.; Chiodini, P. Unsupervised Hierarchical Classification Approach for Imprecise Data in the Breast Cancer Detection. Entropy 2022, 24, 926. https://doi.org/10.3390/e24070926

AMA Style

Fordellone M, Chiodini P. Unsupervised Hierarchical Classification Approach for Imprecise Data in the Breast Cancer Detection. Entropy. 2022; 24(7):926. https://doi.org/10.3390/e24070926

Chicago/Turabian Style

Fordellone, Mario, and Paolo Chiodini. 2022. "Unsupervised Hierarchical Classification Approach for Imprecise Data in the Breast Cancer Detection" Entropy 24, no. 7: 926. https://doi.org/10.3390/e24070926

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop