Next Article in Journal
Optimization Algorithm to Sequence the Management Processes in Information Technology Departments
Next Article in Special Issue
Wavelet Power Spectral Domain Functional Principal Component Analysis for Feature Extraction of Epileptic EEGs
Previous Article in Journal
Synthesis, Mass Spectroscopy Detection, and Density Functional Theory Investigations of the Gd Endohedral Complexes of C82 Fullerenols
Previous Article in Special Issue
Integrated Multi-Model Face Shape and Eye Attributes Identification for Hair Style and Eyelashes Recommendation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Model Approach and Fuzzy Clustering for Mammogram Tumor to Improve Accuracy

1
Department of Mathematics, Indian Institute of Engineering Science and Technology, Shibpur, Howrah 711103, India
2
Institute of Research and Development of Processes, University of the Basque Country, 48940 Leioa, Spain
*
Author to whom correspondence should be addressed.
Computation 2021, 9(5), 59; https://doi.org/10.3390/computation9050059
Submission received: 29 March 2021 / Revised: 7 May 2021 / Accepted: 14 May 2021 / Published: 18 May 2021

Abstract

:
Breast Cancer is one of the most common diseases among women which seriously affect health and threat to life. Presently, mammography is an uttermost important criterion for diagnosing breast cancer. In this work, image of breast cancer mass detection in mammograms with 1024 × 1024 pixels is used as dataset. This work investigates the performance of various approaches on classification techniques. Overall support vector machine (SVM) performs better in terms of log-loss and classification accuracy rate than other underlying models. Therefore, further extensions (i.e., multi-model ensembles method, Fuzzy c-means (FCM) clustering and SVM combination method, and FCM clustering based SVM model) and comparison with SVM have been performed in this work. The segmentation by FCM clustering technique allows one piece of data to belong in two or more clusters. The additional parts are due to the segmented image to enhance the tumor-shape. Simulation provides the accuracy and the area under the ROC curve for mini-MIAS are 91.39 % and 0.964 respectively which give the confirmation of the effectiveness of the proposed algorithm (FCM-based SVM). This method increases the classification accuracy in the case of a malignant tumor. The simulation is based on R-software.

1. Introduction

Cancer is a disease which arises from cells that leave the cell cycle and start to proliferate in an uncontrolled manner and spread into surrounding tissues. In United States, 40,610 women would die due to breast cancer (BC) in 2017 as estimated by the American Cancer Society (according as AMC, i.e., Annual Maintenance Contract; source: Facts and figures 2017–2018) [1]. In the year of 2015, the number of women died of BC is minimum 571 thousand reported by WHO (World Health Organization) (source: Cancer—WHO fact sheets 2017). Globally, a single woman dies per minute because of breast cancer. This death rate is very high because of delaying tumor-diagnosis but early detection of breast cancer and properly therapy can able for decreasing mortality rates. This proliferation could be induced by hormones that are impinging on the breast. Presently, BC is the top cancer in women worldwide, both in the developed and the developing world. Most of the women are diagnosed in late stages due to lack of their awareness and barriers to access to health services. Nearly 7.6 million people died from cancer in 2008. Mortality because of cancer occurs in low and middle income countries in large percentage (about 70 % ) [2,3]. There are various types of breast cancer (BC). The kind of breast cancer depends on which cells in the breast turn into cancer [4].
It is known that benign tumor cells have well-circumscribed, severely round or in elliptical shape with soft edges but usually malignant cells have ambiguous edges and uncontrolled in shapes which indicate about spiculation (source: Cancer—WHO fact sheets 2017 [5]). The existence of spiculation neighbouring a mass usually indicates a malignant criterion and the more homogeneous form points out a benign mass [6]. However, benign lesions can have spiculated patterns (spiculated mass is a lump of tissue with spikes or points on the surface according to oncology) because of overlapping the tissue on the image-projection. The degree of spiculation of the edge of the tumor is an attribute of malignancy considered by different techniques as a feature to classify a mass [7]. The main factors that influence the risk include being a woman and age. Most breast cancers are found in women who are 50 years old or older.
The medical imaging technology, which is known as mammography, is very popular technique for early detection of BC [7]. However, mammograms are difficult to interpret primarily due to the low contrast and arduous process of understanding due to architectural sophistication. The resemblance of tumor intensity to normal celsl on the margin of the tumor which makes it difficult to differentiate between tumors. For tumor recognition and classification, accurate extraction of characteristics of tumor from a mammogram image is uttermost important. Therefore, the selection of effective features for the process of detection and classification is very much needed. Over recent years there are several methods in the application of BC detection and classification with numerous different features. In general, many of these have been involved with statistical and geometrical features to analyse different problems [8]. In this work, we have concerned specifically with encoding the boundary of the tumor by Fourier expansions which are certainly translation invariant.
Microcalcifications are small calcium accumulations (deposits) identified with white tiny specks on a mammogram. These are usually not a result (symptom) of cancer. However, if they appear in certain patterns and cluster together, then they may be a sign (symptom) of precancerous cells or early breast cancer (BC). Three categories of calcifications have been identified [9]: (i) typically benign (ii) intermediate concern (iii) high probability of malignancy.
Breast cancer is characterized by the presence of a mass accumulated or not accumulated by calcifications. There is a possibility of a cyst that is non-cancerous collection of fluid to resemble a mass in the film. The identical intensities of the masses and the normal tissue and similar morphology of the masses and regular breast textures makes it a tedious task to detect masses in comparison with that of calcifications. The shape, size, location, density and margins of the mass are highly beneficial for the radiologist to evaluate the probability of cancer. A majority of the benign masses have compact, well circumscribed margins and slightly circular or elliptical shapes whereas the malignant lesions are characterized by blurred boundaries, irregular appearances and are occasionally enclosed within a radiating pattern of linear speckles. Nevertheless some benign lesions may also possess speculated appearances or blurred peripheries.
The particles’s shape is an uttermost important to determine the history and behavior. It is most important for having a suitable measurement of particles’s shape, so that the changes in shape and the distinctions between the shapes of various populations can be recognised easily. One of the important methods that is used in describing the shape is roundness. Roundness is a measure of sharpness of edges and corners. For implementing roundness, Fourier transformation is used in some cases. The basic idea is to acquire coordinates on the edge that are being measured. For this purpose, the centre point is computed and then all the edge coordinates are transformed into polar coordinates by considering the centre as origin. Then the Fourier expansion coefficients of the vector of radii ( r ( θ )) are determined and the roundness is computed with the help of these coefficients.
In this work, a clear vision of segmentation has been demonstrated by Fuzzy c-means (FCM) clustering technique that allows one piece of data to belong to two or more clusters. The aim of this work is to gather necessary information of the tumor margin. For classifying the breast cancer dataset, a method is developed as combination of support vector machine and Fuzzy c-means clustering. Performance of the results are also compared with other algorithms. The study seeks to investigate the various advanced algorithms which are preferable to LDA and also investigates the method which is most accurate. A comparison has also been made between the performance of the more complex algorithms with LDA. In this work, improvement of the classification modelling framework has been fitted in two distinct levels: (i) a characterise model and (ii) a class-wise model conditional on a character. Five different methods: linear discriminant analysis (LDA), random forests (RF), support vector machines (SVM), neural networks (NNET) and nuclear penalized multinomial regression (NPMR) have been described in detail in Section 5 and Section 6. The results demonstrate that the developments can be gained over LDA.
In Section 2 and Section 3, we discussed a description of database materials. Then the comparison among the performance of the underlying models are discussed in Section 3. Next, in Section 4 and Section 5 we analyzed the proposed work and Section 6 provides details of results with comparison of results. Section 7 and Section 8 explain the extension part related with Section 6, i.e., multi-model ensembles method, the proposed method and the corresponding steps. The results have been interpreted in the context of Breast Cancer and the efficiency of the proposed method has also been demonstrated. Finally, the last section demonstrates the conclusions of this work.

2. Data Description

The data for Breast Cancer is provided at the link: http://peipa.essex.ac.uk/info/mias.html (accessed on 11 December 2012). The Mammographic Image Analysis Society (MIAS) database (digitised at 50 micron pixel edge) has been reduced to 200 micron pixel edge and padded or clipped such as every image is 1024 × 1024 pixels [10]. In the dataset, there are 322 data that present as normal, benign, and malignant tissue. We are only considering the benign and malignant tissue. In R script, all programs of LR, RF and SVM can be performed using ‘ n n e t ’, ‘ r a n d o m F o r e s t ’ and ‘ e 1071 ’ library respectively and ‘ c a r e t ’ package is also used to construct the model using different classification techniques.

3. Mammogram Database

In this work, to know the validation of the results obtained from the proposed method, we have used the dataset of Mammographic Image Analysis Society (MIAS). There are 117 images of benign and malignant tumors considered here. According to background tissue there are three types of the breast images, namely, fatty-glandular, dense and fatty. The mammogram images also make a categorization as follows: (i) normal, (ii) benign and (iii) malignant. The masses are also classified as: (i) architectural distortion (ARCH), (ii) circumscribed (CIRC), (iii) asymmetry (ASYM), (iv) spiculated (SPIC) and lastly (v) other masses (MISC) after considering the tumor’s shape and the attributes of edges into various ways [11]. In the MIAS dataset, the proper position of tumor is pointed out using the coordinates of the tumor radius and the tumor centre that provide the description of the region of interest (i.e., ROI) of the masses. This ROI gives us the detailed information about tumor margin, density and shape. In this work, we are interested with two broad scenarios [12]:
(i) Benign: It is non-life-threatening or non-cancerous tumor. However, it can turn into a cancer status in a few circumstances. Normally, the immunity system segregates such tumors from other cells and can be easily removed from our body. This immune system is called ‘sac’.
(ii) Malignant: It begins with abnormal and uncontrolled growth of cells that can be spread out rapidly or invade neighbouring tissue. In general, the nuclei of the malignant cells are much larger than in normal cells, which may be cause of life-threatening disease in future stages.

4. Fuzzy c-Means Clustering

Fuzzy c-means clustering is a popular technique to recognise the pattern and widely used fuzzy separation approach [2,13]. This method can make groups of pixels into clusters which appeal as fuzzy functions. The pixels can be portions of various clusters and each data have a distinct membership value between [ 0 , 1 ] . We have considered Y = ( Y l , Y 2 , , Y K ) as an K pixels image to be categorized into ς clusters, where 2 ς K . Each cluster is described by its centre and generally the clusters number depends on the properties and intensity value of pixels of breast tissue [14]. In this work, three cluster centres are used empirically due to proper segmentation of mammogram tumors. The technique is an iterative optimization which attempts to find a cluster center in every group to minimize the cost function of the dissimilarity measure denoted and defined as:
J F C M = j = 1 K i = 1 ς μ i j m y j υ i 2
where y j be the intensity related with the jth pixel, υ i be the center of the ith cluster and the Euclidean distance between pixel y j and cluster center υ i is defined as y j υ i and m be the parameter of weighting exponent on every membership [13]. In the Fuzzy c-means (FCM) method, the membership of pixel y j to the ith cluster is referred as μ i j . The constraints of μ i j are given below:
i = 1 ς μ i j = 1
where
0 μ i j 1
0 < j = 1 K μ i j < K
If the cluster centre has high membership value, proposed by the intensity of pixel, the values of cost function is decreasing and vice versa. A pixel is a member or not of a specific cluster, measured by the presence of membership function. The distance between each cluster center and the pixel is relied by this possibility in FCM method. The membership function and cluster centres in FCM are as follows:
μ k j = 1 k = 1 ς ( y j υ k y j υ i ) 2 m 1
and
υ i = j = 1 K μ i j m x j j = 1 K μ i j m
At first each cluster center is initialized randomly during execution and an iterative method is applied to (5) and (6) for obtaining the cluster center and the membership function [13]. After some iterative steps of the cluster center or the changes in the case of membership function can identify the convergence of this scheme.

5. Some Well-Known Algorithms and Preliminaries

In this work, we demonstrated five different approaches for determining which method is best to classify based on the character and class of the image of cancer cell. This work suggests which complex technique gives the highest predictive accuracy in image of breast cancer cell identification. Let ζ i and ς i be the character and class of the ith cancer-cell, respectively, with ζ i ( 1 , 2 , , K ζ ) and ς i ( 1 , 2 , , K ς ). Furthermore, let ζ = ( ζ 1 , ζ 2 , , ζ η ) and ς = ( ς 1 , ς 2 , , ς η ) so that both ζ and ς are vectors of length η . Additionally, for the ith cancer-cell, let the vector A i = ( a i 1 , b i 1 , c i 1 , d i 1 , a i 2 , b i 2 , , a i H , b i H , c i H , d i H ), which will have length 4 H , where H is the number of harmonics. Here, H = 15 is used. Let A = ( A 1 , A 2 , , A η ) be matrix that will contain η rows and 4 H columns. Finally, let A = c o v ( A ) and using spectral decomposition we have, A = P Λ 1 2 Λ 1 2 P , where Λ is a diagonal matrix whose elements are the eigenvalues of A and P is a matrix of eigen-vectors of A [15]. Now let, Y = A Γ , where Y is an η by 4 H matrix of the coordinates rotated by the principal components and Γ = Λ 1 2 P . The matrix A and the first p columns of Y are used as potential predictors in each underlying techniques. The augmented matrix is denoted X = ( A : Y ) , where X is the augmented matrix combining A and Y that contains η rows and 4 H + p columns, where p 4 H . The procedures are described here. The model is as follows:
p i κ ( ζ ) = P ( ζ i = κ | X ) , i = 1 , , η
These probabilities are estimated by the various methods and a series of models were fitted for class-wise classification conditional on each character. Formally,
p i g ( ς | ζ = κ ) = P ( ς i = g | ζ i = κ , X ) , for i = 1 , , η
This conditional probability is computed as:
p i g κ ( ς , ζ ) = P ( ς i = g , ζ i = κ | X ) = P ( ς i = g | ζ i = κ , X ) P ( ζ i = κ | X )
One convenient property of structuring the models in this manner is that
g ς ( t ) P ( ς i = g , ζ i = κ | X ) = P ( ζ i = κ | X )
where ς ( t ) is the set of classes that fall into character ζ = κ . This means whenever estimating the probability that a particular breast cancer-cell belongs to a particular character, the implied probability that a breast cancer-cell belongs to a particular character that can be calculated by summing across all of the class that are nested within a particular character and this probability will be consistent with probability calculated from the character model alone. This entire modeling process was repeated separately for the six different criteria considered (e.g., Centre x a , Centre y a , etc.). Every method is described in terms of the character classification model for illustration purposes. The methods are easily modified for classifying class within a given character.

5.1. Linear Discriminant Analysis (LDA)

The assumption of this method is that the distributions of the features within every specific character of background tissue follow the multivariate Gaussian densities with common covariance matrix L D across all of the N ζ characters [16]. Let p κ be the prior probability of belonging to character k, then the discriminant function is denoted and defined as:
δ κ ( x ) = x ζ Σ 1 μ κ 1 2 μ κ ζ + log ( p κ )
The parameters are unknown which must be computed such as:
p ^ κ = η κ η
μ ^ κ = i = 1 η κ x i η κ
Σ ^ = κ = 1 K ζ i = 1 η κ ( x i μ ^ κ ) ( x i μ ^ κ ) η η κ
where η κ denotes the number of observations in the κ t h character with η = κ = 1 K ζ η κ . Every observation is classified as [16,17]:
ζ ^ i = argmax κ δ κ ( x i )

5.2. Random Forest (RF) Algorithm

Random forest (RF), a decision tree-based regression and classification approach, is widely used and most flexible [18,19] and this algorithm also establishes multitude decision trees using Bootstrap [20]. In this work, the algorithm (tree model) tries to classify every observation into one of K ζ categories. This model comprises partitioning the feature space into M regions (i.e., R 1 , R 2 , , R M ) and
p m κ ^ = 1 η m x i R m I ( ζ i = κ )
where m be the node index of region R m , containing N m observations and m = 1 M η m = η . The model can classify (as the majority class in node m) the observations which fall into node m and the Gini index (the measure of node impurity) as follows:
G m = κ κ p ^ m κ p ^ m κ = κ = 1 K ζ p ^ m κ ( 1 p ^ m κ )
Here we considered a subset of the variables as possible splits that can reduce the correlation among the trees beyond bootstrapping alone at every potential split in all trees [19].

5.3. Support Vector Machine (SVM)

Support vector machine is a new classification algorithm for development and implementation which is currently of great interest in theory and application research purpose in distinct fields [21]. When constraints are imposed, this mechanism involves for optimizing a convex loss function, unaffected by local minima problem. The diagram of SVM is depicted in Figure 1. A support vector classifier obtains a hyper plane such as:
{ x : f ( x ) = x T β + β 0 = 0 }
Often, the derivation is made for more than two classes, i.e., in the case of multi-class purpose. For classification purpose, X η , a random η -vector is chosen. Let, Z { 1 , 2 , 3 , , K } is label of class and K is denoted as the number of classes. There are various methods to interpret such a multi-class SVM strategy. This method divides the K-classes of problems into κ 2 comparisons, for all pairs of classes. A classifier f ϱ κ ^ is built such as: the ϱ th class considered as positive and the kth class considered as negative and ϱ , k = 1 , 2 , , K where ϱ k . Multi-class classification is successfully gained by constructing and combining several binary classifiers. Whenever the data-set is not linearly separable, a kernel function is involved to transform the data-set from the input space to a higher-dimensional output space in the input space where the data can be linearly separable [22]. A cost parameter is used for minimizing the misclassifications number.

5.4. Nuclear Penalized Multinomial Regression (NPMR)

In multinomial regression model, it is assumed that the distribution of characters follows a multinomial distribution which is conditional on independent variables [23]. This regression model is defined as a series consisting of transformations of logit function as follows:
log P ( ζ = 1 | X = x ) P ( ζ = K ζ | X = x ) = θ 1 + β 1 X log P ( ζ = 2 | X = x ) P ( ζ = K ζ | X = x ) = θ 2 + β 2 X log P ( ζ = K ζ 1 | X = x ) P ( ζ = K ζ | X = x ) = θ K ζ 1 + β K ζ 1 X
In the present study, each β k is assumed to be a vector with a length of 4 H + p . There are a total of ( K ζ 1 ) of these vectors. These vectors can then be combined into a matrix denoted by Λ having dimension ( 4 H + p ) × ( K ζ 1 ) . Then the log-likelihood L ( θ , Λ ; X , ζ ) can be defined by the use of these equations.
max θ R K ζ 1 , Λ R ( 4 H + p ) × ( K ζ 1 ) L ( θ , Λ ; X , ζ )
The maximum likelihood estimate of the parameters is obtained by ( 5 ) .
In NPMR, the following set-up is assumed:
max θ R K ζ 1 , Λ R ( 4 H + p ) × ( K ζ 1 ) L ( θ , Λ ; X , ζ ) λ Λ 0 , where Λ 0 = r = 1 r κ ( Λ ) σ r
Here σ r are the singular values of Λ . Optimal values of λ can be found by cross validation.

5.5. Neural Networks (NNET)

Neural networks is considered as two-stage classification methods [24]:
(i)
At first, in the matrix X, the input features, transform into a matrix of derived features Z whose dimension is η × Q .
(ii)
Every character is modelled as a function of Z in the following manner:
Z q = σ ( ζ 0 q + ζ q X ) ,     where q = 1 , , Q
U κ = β 0 κ + β κ Z ,     where κ = 1 , , K ζ
f κ ( X ) = h κ ( U ) ,     where t = 1 , , K ζ
σ is defined as: σ ( η ) = 1 / ( 1 + e η ) . From the output U κ , the function h permits for a transformation such as h κ ( U ) = e U κ j = 1 K ζ e U j . This function is same transformation which is used in multinomial regression models. Using cross-validation, tuning parameters of underlying model are perceived.

6. Results and Comparison

For evaluating the accuracy for each classification technique of each image of cancer cell for both character and class, cross-validation (six-fold) is used. Each method has been evaluated using the criteria of log loss and classification accuracy. In the case of character, log-loss was computed as follows:
logLoss C h a r a c t e r = i = 1 η κ = 1 K ζ I ( ζ i = κ ) log ( p ^ i κ ( ζ ) )
where p ^ i ( ζ ) is the vector having length K ζ . This vector contains the predicted probabilities of each of the K ζ characters for the i t h observation and
I ( ζ i = κ ) = 1 , if ζ i = κ 0 , otherwise
For the purpose of class, log-loss is defined as follows:
logLoss C l a s s = i = 1 η g = 1 K ς I ( ς i = g ) log ( p ^ i g ( ς , ζ ) )
where the vector p ^ i g ( ς , ζ ) with length K ς is containing the predicted probabilities of each of the K ς classes for the i t h observation. Here
I ( ς i = g ) = 1 , if ς i = g 0 , elsewhere
Log loss has been chosen because of the fact that it heavily penalizes a technique for being over-confident and incorrect.
When we predict character in terms of log loss, the method SVM is the best model (ranged from 0.27 through 0.34) among all underlying models. In this work, NNET is the second best among all. The method LDA is the worst among all the performing models (since values of log-loss ranging from 1.06 to 1.65). Comparison among log-loss results for the purpose of character is also performed (shown in Table 1). Next, the results have been provided for class purpose models in terms of log-loss and it is predicted that SVM method is also the best among all with log-loss values (range from 0.72 through 0.98). Whenever we are predicting classes, there does not present any clear view for choosing second best among rest underlying models. Here also LDA provides worst results (since the values of log-loss ranging between 2.82 to 3.53).
Furthermore, classification accuracy is measured for different approaches for evaluating each model. For observation i, each probability vector predicting breast cancer cell is classified as belonging to the character and class with the highest predicted probability in the vector and the results also have been compared in this work.

Accuracy Results

Accuracy C h a r a c t e r = i = 1 η I ( ζ i = ζ ^ i ) η , where ζ ^ i = { κ : max κ p ^ i κ ( ζ ) }
Accuracy C l a s s = i = 1 η I ( ς i = ς ^ i ) η , where ς ^ i = { g : max g p ^ i g κ ( ς , ζ ) }
The measurement is very simple percentage form of breast cancer cell which are classified to the true character or class.
In terms of accuracy for the character and class models there is a clear vision that overall percentage of SVM is the best among all except in the case of centre of x co-ordinates during characterised model performance (shown in Table 2). In the class based modelling the best performing method is SVM as well as NNET (though it is never the best in the characterised modelling). Finally, it is noted that LDA is worst among all for both cases with respect to accuracy (shown in Table 2). As per the obtained results, SVM is the best among all, so we proceed with this method for classification as benign or malignant tumor criterion with multi approaches. Next, we also proceed with a parallel combination of FCM clustering and SVM as unsupervised and supervised algorithms respectively.

7. Classifier Ensembles

A multi-model approach can obtain a more accurate prediction and significantly improve prediction. The approach provides better performance than the single-model approach in terms of statistical accuracy, without unduly increasing the computational burden [25]. Therefore, in this work, we use MIAS datasets to demonstrate the multi-model method which can provide comparable prediction performance. This approach is used for combining the predictive ability of multiple models for better prediction accuracy. The dataset is partitioned into training, development and test sets and build a series of models. Multi-model approach has achieved high accuracy on nuclear segmentation and classification after correctly resolving ambiguities in clustered regions containing heterogeneous cell populations [26]. There are mainly two techniques which are widely used for combining multiple classifiers, i.e., bagging and boosting. In bagging, several classifiers are trained independently by different training sets through the bootstrap method [27]. Bootstrapping builds k replicate training data sets to construct k independent classifiers by randomly re-sampling the original given training dataset with replacement. Then the k classifiers are aggregated through an appropriate combination method, such as majority voting [28]. Each classifier is trained using a distinct training set in boosting. However, the k classifiers are trained not in parallel and independently, but sequentially. Schapire (1990) proposed the original boosting approach, boosting by filtering. AdaBoost (or, Adaptive Boosting) is also the common boosting used in pattern recognition [29].
For evaluating the performance of mammographic tumor classification, accuracy are being measured by SVM and SVM ensembles and the results (from Table 3) show that the performance of SVM ensembles is better than SVM, based on MIAS datasets prediction. Additionally in terms of accuracy prediction, SVM ensembles based on bagging provide better performance than other classifiers in this work.

8. Fundamental Evaluation Measures

In classification purpose, generally the classifier is evaluated by a confusion matrix. For a binary classification problem a matrix is a square of 2 × 2 as shown in Table 4. In this table, column represents the classifier prediction, while the row is the real value of class label.
The acronym TP, FN, FP, and TN of the confusion matrix cells refers as follows:
TP:
true positive (the number of positive cases that are correctly identified as positive),
FN:
false negative (the number of positive cases that are misclassified as negative cases),
FP:
false positive (the number of negative cases that are incorrectly identified as positive cases),
TN:
true negative (the number of negative cases that are correctly identified as negative cases).
Accuracy, the most common metric for classifier evaluation, it assesses the overall effectiveness of the algorithm by estimating the probability of the true value of the class label. Accuracy is defined as:
T P + T N T P + T N + F P + F N
Furthermore, error rate is an estimation of misclassification probability according to model prediction, which is defined as follows:
F P + F N T P + T N + F P + F N
Error, an instance class, is predicted incorrectly where error rate be a percentage of errors made over the whole set of instances used for testing purpose.

8.1. FCM and SVM Combination Method

This method is a combination of two techniques: (i) Fuzzy c-means and (ii) support vector machine. The classification is investigated only in the case of cluster centres. This is achieved as a result of fuzzy clustering as described in Section 4 which is used for the purpose of training sample of SVM method shown in Figure 2 step by step.

8.2. FCM Based on SVM Model

In this work, a parallel combination is made using Fuzzy c-means (FCM) and support vector machine (SVM). Individual clustering was performed for each class. For this method, FCM clustering is used to pick up instances represented of class of breast cancer cell for each cancer cell in original training data-set. Original data and centres of cluster are also being obtained from FCM that are used as training to SVMs shown in Figure 3 step by step. It is a linear combination of basis functions. The following section describes FCM-based SVM structure step by step. The steps are as follows:
Step 1: In first step, no computation is performed. Each node in this step corresponds to one input variable and directly transmits input values x to the next step. When the FCM-based SVM is used as an equaliser, the input variables are x in the equation x ( k ) = [ x ( k ) , x ( k 1 ) , , x ( k n + 1 ) ] , where x ( k ) be the output.
Step 2: Each node in this step represents a basis function and may be regarded as a cluster. The number of nodes in this step is equal to the number of clusters in FCM. The output of each node is a membership value that specifies the degree to which an input value belongs to a fuzzy cluster. Let, r be the number of nodes (clusters) in the step and let v 1 ,..., v r denote the centres of these r clusters. According to FCM [30], the membership value calculated by each basis function (node) is as follows:
u i ( x ) = ( 1 x v i 2 ) 1 m 1 j = 1 r ( 1 x v j 2 ) 1 m 1
where m is the parameter of weighting exponent on every membership [13].
Step 3: Each node in this step corresponds to one output node. Each node performs a weighted sum of membership values sent from nodes in step 2.

8.3. Performance Evaluation and Discussion

For evaluating the performance of mammographic tumor classification, receiver operating characteristics (ROC) curves and accuracy are measured by the proposed schemes. The ROC curve indicates the main contrast capacity of various methods. Apart from this, accuracy can also be measured for necessary significant evaluation. Accuracy is uttermost important because it provides precise rate of classification of two different levels (i.e., benign and malignant). It should be as high as possible for a classification model.
In Figure 4, the performance of different classifiers with the MIAS datasets is depicted with ROC plots using ‘ROCR’ package in R-software. The AUC can measure the performance of classification which is between 0 and 1. It summarizes the entire location of the ROC curve rather than depending on a specific operating point [31,32]. For FCM-based SVM method, the AUC is 0.964 which indicates that it provides the best performance among all (from Figure 4). The accuracy of this method is 91.39 % . The reduction of error rate is compared with each of the proposed models. The performance of the model by ROC curve is also measured using estimation of the Area Under Curve (AUC) value.
AUC (based on ROC) and accuracy seems not the same concept. Accuracy is based on one specific cutpoint, while ROC tries to cover all the cutpoint and plots the sensitivity and specificity. Whenever we compare the overall accuracy, we are comparing the accuracy based on some cutpoint. The overall accuracy varies from different cutpoint. The area under the curve (AUC) is equal to the probability that a classifier will rank a randomly chosen positive instance higher than a randomly chosen negative example. It measures the classifiers skill in ranking a set of patterns according to the degree to which they belong to the positive class, but without actually assigning patterns to classes. The accuracy also depends on the ability of the classifier to rank patterns, but also on its ability to select a threshold in the ranking used to assign patterns to the positive class if above the threshold and to the negative class if below. Thus the classifier with the higher AUC is likely to also have a higher accuracy as the ranking of patterns which is beneficial to both AUC and accuracy. However, if one classifier ranks patterns well, but selects the threshold badly, it can have a high AUC but a poor accuracy. The high value of AUC indicates how strongly any predictor is capable for classification between two different classes of the response variable. Features with higher AUC are selected and overall performance of these features have been estimated by constructing a model using RF. It is observed that FCM based on SVM classification can maximize both AUC and accuracy.
The changing structure of cells can be recognized by analyzing the images obtained from mammogram method. Such images are very complex in nature and need fruitful information to analyse the malignancy properly. FCM method is used mainly for segmentation image to enhance the tumor’s shape. Although the SVM classification for breast cancer detection has achieved with high accuracy, but with FCM-based SVM method the accuracy rate is increased which provides that the rate of proposed method (FCM-based SVM) is higher than all previous methods. The success rate of SVM is 87.258 % but with the FCM-based SVM method, the success rates are increased to 91.39 % , shown in Table 5. After comparison of all underlying methods (described in Section 7 and Section 8), FCM-based SVM gives better accuracy than others proposed methods.

9. Concluding Remarks

Breast cancer (BC) is a very serious disease among women worldwide as it is accountable for increasing death rates among women. For improving the current era with BC is a vital concern and it can be achieved by fruitful diagnosis, proper investigation, appropriate patient, and with proper clinical management. To identify BC in the early stages, a regular basic check-up ca save many lives of human beings. The status of this cancer is changing with time since the distribution, appearance, and structural geometry of the cells change on a particular time basis due to the chemical changes which are always going on inside the cells. Due to the non trivial nature of the images the physician sometimes making a decision, might counter others. However, computer-aided-diagnosis methods can be enlightened a significant amount of information from the obtained images and also provide a proper decision based on the achieved information (viz., identification of cancer) using classifications of the images. Here, five various approaches including LDA were compared based on their performance in classifying the characters and classes of image data-set of breast cancer cells. First we demonstrated that classification accuracy can be increased in terms of both rate of misclassification and log-loss using SVM method which performs very well and slightly out performed NNET. Both (i.e., SVM and NNET) performed much better than RF, NPMR, and LDA.
The contribution of these methods is of utmost importance and using such advanced techniques with proposed modifications are used for image classification purpose, especially for image of breast cancer classification and also for segmentation. Overall SVM performs better than other underlying models, so additional extension has been taken into account in this current work with classifier ensembles. Then the further utilizations of FCM and SVM combination method and FCM based on the SVM method make a delightful scenario for the image classification. The comparison results among them provide higher accuracy, such enhancement confirms the effectiveness of the proposed approach for malignant tumor with the goal of increasing the accuracy.

Author Contributions

Conceptualization, S.G., G.S. and M.D.l.S.; Methodology, S.G., G.S. and M.D.l.S.; Investigation, S.G., G.S. and M.D.l.S.; Formal analysis, S.G., G.S. and M.D.l.S.; Writing—original draft preparation, S.G., G.S. and M.D.l.S.; Writing—review and editing, S.G., G.S. and M.D.l.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Spanish Government for its support through grant RTI2018-094336-B-100 (MCIU/AEI/FEDER, UE) and to the Basque Government for its support through grant IT1207-19.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data used to support the findings of this study are included in the references within the article.

Acknowledgments

The authors are grateful to the anonymous referees and Editors for their careful reading, valuable comments and helpful suggestions, which have helped them to improve the presentation of this work significantly. The third author (Manuel De la Sen) is grateful to the Spanish Government for its support through grant RTI2018-094336-B-100 (MCIU/AEI/FEDER, UE) and to the Basque Government for its support through grant IT1207-19.

Conflicts of Interest

The authors declare that they have no conflict of interest regarding this work.

References

  1. Facts and Figures. 2017–2018. Available online: https://www.cancer.org/research/cancer-facts-statistics8 (accessed on 5 March 2018).
  2. Bezdek, J.C. Pattern Recognition with Fuzzy Objective Function Algorithms; Springer Science and Business Media: Berlin, Germany, 2013. [Google Scholar]
  3. Ghosh, S.; Samanta, G.P. Statistical modelling for cancer mortality. In Letters in Biomathemtics; Taylor and Francies: Abingdon, UK, 2019; Volume 6. [Google Scholar] [CrossRef] [Green Version]
  4. Howlader, N.; Noone, A.M.; Krapcho, M. SEER Cancer Statistics Review, 1975–2014; National Cancer Institute: Bethesda, MD, USA, 2017. [Google Scholar]
  5. Cancer—WHO Fact Sheets. 2017. Available online: http://www.who.int/mediacentre/factsheets/fs297/en/ (accessed on 14 October 2017).
  6. Sickles, E.A. Breast masses: Mammographic evaluation. Radiology 1989, 173, 297–330. [Google Scholar] [CrossRef] [PubMed]
  7. Mohanty, F.; Rup, S.; Dash, B.; Majhi, B.; Swamy, M. Mammogram classification using contourlet features with forest optimization-based feature selection approach. Multimed. Tools Appl. 2018, 78, 12805–12834. [Google Scholar] [CrossRef]
  8. Cawley, G.C.; Talbot, N.L. On over-fitting in model selection and subsequent selection bias in performance evaluation. J. Mach. Learn. Res. 2010, 11, 2079–2107. [Google Scholar]
  9. Gorgel, P.; Sertbas, A.; Uçan, O.N. Computer-aided classification of breast masses in mammogram images based on spherical wavelet transform and support vector machines. Expert Syst. 2015, 32, 155–164. [Google Scholar] [CrossRef]
  10. MIAS Database. 2012. Available online: http://peipa.essex.ac.uk/info/mias.html (accessed on 11 December 2012).
  11. Feig, S.A.; Yaffe, M.J. Digital mammography, computer-aided diagnosis, and telemammography. Radiol. Clin. N. Am. 1995, 33, 1205–1230. [Google Scholar] [PubMed]
  12. James, A.P.; Sugathan, S. Parallel Realization of Cognitive Cells on Film Mammography. In Proceedings of the 12th IEEE International Conference on Trust, Security and Privacy in Computing and Communications, Melbourne, Australia, 16–18 July 2013. [Google Scholar]
  13. Chuang, K.S.; Tzeng, H.L.; Chen, S.; Wu, J.; Chen, T.J. Fuzzy c-means clustering with spatial information for image segmentation computerized medical imaging and graphics. Comput. Med. Imaging Graph. 2006, 30, 9–15. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  14. Keller, B.; Nathan, D.; Wang, Y.; Zheng, Y.; Gee, J.; Conant, E.; Kontos, D. Adaptive multi-cluster fuzzy c means segmentation of breast parenchymal tissue in digital mammography. In International Conference on Medical Image Computing and Computer-Assisted Intervention; Springer: Toronto, ON, Canada, 2011; pp. 562–569. [Google Scholar]
  15. Johnson, R.A.; Wichern, D. Applied Multivariate Statistical Analysis, 6th ed.; Prentice Hall: Englewood Cliffs, NJ, USA, 2002. [Google Scholar]
  16. Fisher, R.A. The use of multiple measurements in taxonomic problems. Ann. Eugen. 1936, 7, 179–188. [Google Scholar] [CrossRef]
  17. Hastie, T.; Tibshirani, R.; Friedman, J. The Elements of Statistical Learning: Data Mining, Inference and Prediction, 1st ed.; Springer: New York, NY, USA, 2001. [Google Scholar]
  18. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef] [Green Version]
  19. Everson, T.M.; Lyons, G.; Zhang, H.; Soto-Ramirez, N.; Lockett, G.A.; Patil, V.K.; Merid, S.K.; Soderhall, C.; Melen, E.; Holloway, J.W.; et al. DNA methylation loci associated with atopy and high serum IgE: A genome-wide application of recursive Random Forest feature selection. Genome Med. 2015, 7, 89. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  20. Efron, B. Bootstrap methods: Another look at the jackknife. Ann. Stat. 1979, 7, 1–26. [Google Scholar] [CrossRef]
  21. Boser, B.E.; Guyon, I.M.; Vapnik, V.N. A training algorithm for optimal margin classifiers. In Proceedings of the 5th Annual ACM Workshop on Computational Learning Theory, Pittsburgh, PA, USA, 27–29 July 1992; ACM Press: New York, NY, USA, 1992; pp. 144–152. [Google Scholar]
  22. Chang, C.; Lin, C. LIBSVM: A library for support vector machine. ACM Trans. Intell. Syst. Technol. 2011, 2. [Google Scholar] [CrossRef]
  23. Venables, W.N.; Ripley, B.D. Modern Applied Statistics with S, 4th ed.; Springer: New York, NY, USA, 2002. [Google Scholar]
  24. Powers, S.; Hastie, T.; Tibshirani, R. Nuclear penalized multinomial regression with an application to predicting at bat outcomes in baseball. Stat. Model. 2016, 18, 388–410. [Google Scholar] [CrossRef] [PubMed]
  25. Stoica, P.; Selen, Y.; Li, J. Multi-model approach to model selection. Digit. Signal Process. 2004, 14, 399–412. [Google Scholar] [CrossRef]
  26. Lin, G.; Chawla, M.K.; Olson, K.; Barnes, C.A.; Guzowski, J.F.; Bjornsson, C.; Shain, W.; Roysam, B. A Multi-Model Approach to Simultaneous Segmentation and Classification of Heterogeneous Populations of Cell Nuclei in 3D Confocal Microscope Images. Cytom. Part A 2007, 71, 724–736. [Google Scholar] [CrossRef] [PubMed]
  27. Breiman, L. Bagging predictors. Mach. Learn. 1996, 24, 123–140. [Google Scholar] [CrossRef] [Green Version]
  28. Kim, H.C.; Pang, S.; Je, H.M.; Kim, D.; Bang, S.Y. Constructing support vector machine ensemble. Pattern Recognit. 2003, 36, 2757–2767. [Google Scholar] [CrossRef]
  29. Schapire, R.E. The strength of weak learnabilty. Mach. Learn. 1990, 5, 197–227. [Google Scholar] [CrossRef] [Green Version]
  30. Hoppner, F.; Klawonn, F.; Kruse, R.; Runkler, T. Fuzzy Cluster Analysis: Methods for Classification, Data Analysis and Image Recognition; Wiley: New York, NY, USA, 1999. [Google Scholar]
  31. Hanley, J.A.; McNeil, B.J. The meaning and use of the area under a receiver operating characteristic (ROC) curve. Radiology 1982, 143, 29–36. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  32. Swets, J.A. ROC analysis applied to the evaluation of medical imaging techniques. Investig. Radiol. 1979, 14, 109–121. [Google Scholar] [CrossRef]
Figure 1. Structure of support vector machine.
Figure 1. Structure of support vector machine.
Computation 09 00059 g001
Figure 2. Block scheme for FCM and SVM combinatione.
Figure 2. Block scheme for FCM and SVM combinatione.
Computation 09 00059 g002
Figure 3. Block scheme for FCM-based SVM.
Figure 3. Block scheme for FCM-based SVM.
Computation 09 00059 g003
Figure 4. ROC curve.
Figure 4. ROC curve.
Computation 09 00059 g004
Table 1. Logloss comparison.
Table 1. Logloss comparison.
Character Class
LDARFSVMNPMRNNET LDARFSVMNPMRNNET
Centre x B 1.3410.3310.3290.3570.343 3.4500.8450.7960.8020.811
Centre y B 1.6530.3250.3210.3280.321 3.4540.7870.7280.8840.741
Radius B 1.0470.3350.2710.3350.326 3.3530.8290.8290.810.829
Centre x M 1.0120.3320.3380.3550.330 3.2790.9970.9880.9921.002
Centre y M 1.2710.3230.3180.3730.363 2.8220.9770.9820.9851.002
Radius M 1.0930.3270.2980.3480.320 3.1030.9760.9570.9591.003
Table 2. Accuracy comparison.
Table 2. Accuracy comparison.
Character Class
LDARFSVMNPMRNNET LDARFSVMNPMRNNET
Centre x B 0.8120.8330.8980.8570.877 0.7450.7890.8960.8890.895
Centre y B 0.7780.8990.9360.9080.901 0.6540.7990.8290.8180.829
Radius B 0.7470.8450.8850.8350.826 0.6180.7800.7850.7810.784
Centre x M 0.8820.8890.9210.8910.885 0.7790.8350.8510.8330.849
Centre y M 0.8280.8430.8950.8790.863 0.5820.8660.8820.8500.882
Radius M 0.8210.8390.8410.8430.820 0.7010.8530.8570.8490.857
Table 3. Classification accuracies of single SVM and SVM ensembles.
Table 3. Classification accuracies of single SVM and SVM ensembles.
ClassificationMethodsAccuracy in Percentage
SVMSingle87.258
Bagging89.323
Boosting88.575
Table 4. Confusion matrix for two classes classification.
Table 4. Confusion matrix for two classes classification.
Predicted PositivePredicted Negative
Actual positiveTPFN
Actual negativeFPTN
Table 5. Performance evaluation and comparisonn.
Table 5. Performance evaluation and comparisonn.
MethodsAccuracy in PercentageError Rate
SVM87.2580.127
FCM & SVM82.330.177
FCM based SVM91.390.086
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Ghosh, S.; Samanta, G.; De la Sen, M. Multi-Model Approach and Fuzzy Clustering for Mammogram Tumor to Improve Accuracy. Computation 2021, 9, 59. https://doi.org/10.3390/computation9050059

AMA Style

Ghosh S, Samanta G, De la Sen M. Multi-Model Approach and Fuzzy Clustering for Mammogram Tumor to Improve Accuracy. Computation. 2021; 9(5):59. https://doi.org/10.3390/computation9050059

Chicago/Turabian Style

Ghosh, Sarada, Guruprasad Samanta, and Manuel De la Sen. 2021. "Multi-Model Approach and Fuzzy Clustering for Mammogram Tumor to Improve Accuracy" Computation 9, no. 5: 59. https://doi.org/10.3390/computation9050059

APA Style

Ghosh, S., Samanta, G., & De la Sen, M. (2021). Multi-Model Approach and Fuzzy Clustering for Mammogram Tumor to Improve Accuracy. Computation, 9(5), 59. https://doi.org/10.3390/computation9050059

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop