Next Article in Journal
On the Diagnosis of Aortic Dissection with Impedance Cardiography: A Bayesian Feasibility Study Framework with Multi-Fidelity Simulation Data
Previous Article in Journal
Intracellular Background Estimation for Quantitative Fluorescence Microscopy
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Proceeding Paper

A Complete Classification and Clustering Model to Account for Continuous and Categorical Data in Presence of Missing Values and Outliers

by
Guillaume Revillon
and
Ali Mohammad-Djafari
*
L2S, CentraleSupélec-Univ Paris Saclay, 91192 Gif-sur-Yvette, France
*
Author to whom correspondence should be addressed.
Presented at the 39th International Workshop on Bayesian Inference and Maximum Entropy Methods in Science and Engineering, Garching, Germany, 30 June–5 July 2019.
These authors contributed equally to this work.
Proceedings 2019, 33(1), 23; https://doi.org/10.3390/proceedings2019033023
Published: 9 December 2019

Abstract

:
Classification and clustering problems are closely connected with pattern recognition where many general algorithms have been developed and used in various fields. Depending on the complexity of patterns in data, classification and clustering procedures should take into consideration both continuous and categorical data which can be partially missing and erroneous due to mismeasurements and human errors. However, most algorithms cannot handle missing data and imputation methods are required to generate data to use them. Hence, the main objective of this work is to define a classification and clustering framework that handles both outliers and missing values. Here, an approach based on mixture models is preferred since mixture models provide a mathematically based, flexible and meaningful framework for the wide variety of classification and clustering requirements. More precisely, a scale mixture of Normal distributions is updated to handle outliers and missing data issues for any types of data. Then a variational Bayesian inference is used to find approximate posterior distributions of parameters and to provide a lower bound on the model log evidence used as a criterion for selecting the number of clusters. Eventually, experiments are carried out to exhibit the effectiveness of the proposed model through an application in Electronic Warfare.

1. Introduction

Classification and clustering problems are closely connected with pattern recognition [1] where many general algorithms [2,3,4] have been developed and used in various fields [5,6]. Depending on the complexity of patterns in data, classification and clustering procedures should take into consideration both continuous and categorical data which can be partially missing and erroneous due to mismeasurements and human errors. However, most algorithms cannot handle missing data and imputation methods [7] are required to generate data to use them. Hence, the main objective of this work is to define a classification and clustering framework that handles both outliers and missing values. Here, an approach based on mixture models is preferred since mixture models provide a mathematically based, flexible and meaningful framework for the wide variety of classification and clustering requirements [8]. Two families of models emerge from finite mixture models fitting mixed-type data:
  • The location mixture model [9] that assumes that continuous variables follow a multivariate Gaussian distribution conditionally on both component and categorical variables.
  • The underlying variables mixture model [10] that assumes that each discrete variable arises from a latent continuous variable and that all continuous variables follow a Gaussian mixture model.
In this work, the location mixture model approach is retained since it better models relations between continuous and categorical features when data patterns are mostly designed by first choosing patterns of categorical features to achieve a specific goal and then choosing continuous features that meet constraints related to the chosen patterns and the problem environment. Indeed regarding clustering approach, each cluster groups observations that share same combinations of categorical features where continuous features belong to a peculiar subset. Hence, the location mixture model naturally responds to that dependence structure by assuming that continuous variables are normally distributed conditionally to categorical variables. More precisely, a scale mixture of conditional Gaussian distributions [11] is updated to handle outliers and missing data issues for any types of data. Then a variational Bayesian inference [12] is used to find approximate posterior distributions of parameters and to provide a lower bound on the model log evidence used as a criterion for selecting the number of clusters. An application of the resulting model in Electronic Warfare [13] is proposed to perform Source Emission Identification which is a supreme asset for decision making in military tactical situations. By providing information about the presence of threats, classification and clustering of radar emitters have a significant role ensuring that countermeasures against enemies are well-chosen and enabling detection of unknown radar signals to update databases. As a pulse-to-pulse modulation pattern [14], a radar signal pattern is decomposed into a relevant arrangement of sequences of pulses where each pulse is defined by continuous features and each sequence is characterized by categorical features. However, a radar signal is often partially observed due to the presence of many radar emitters in the electromagnetic environment causing mismeasurements and measurement errors. Therefore the proposed model is suitable for radar emitter classification and clustering. The outline of the paper is as follows. Assumptions on mixed-type data are presented in Section 2. Then, the proposed model and inference procedure are introduced in Section 3. Finally, evaluation of the model is proposed through different experiments on radar emitter datasets in Section 4.

2. Mixed-Type Data

In this section, a joint distribution for mixed data is introduced to model the dependence structure between continuous and categorical data. Then, outliers and missing values are tackled by taking advantage of the joint distribution.

2.1. Assumptions on Mixed-Type Data

Data x consist of J observations ( x j ) j = 1 J gathering continuous features x q = ( x q j ) j = 1 J and categorical features x c = ( x c j ) j = 1 J . Let x j = ( x q j , x c j ) the j t h observation vector of mixed variables where
  • x q j R d is a vector of d continuous variables,
  • x c j = x c j 0 , , x c j q 1 C q is a vector of q categorical variables where C q = C 0 × × C q 1 is the tensor gathering each space C i = m 1 i , , m | C i | i of events that x c j i can take i { 0 , , q 1 } .

2.2. Distribution of Mixed-Type Data

Considering that the retained approach focuses on conditioning continuous data x q = ( x q j ) j = 1 J according to categorical data x c = ( x c j ) j = 1 J , the following joint distribution is introduced
j { 1 , , J } , p ( x q j , x c j ) = c C q π c N x q j | μ c , Σ δ x c j c
where continuous variables x q j are normally distributed according to categorical variables x c j with means ( μ c ) c C q and variance Σ . As for categorical variables x c j , they are jointly distributed according to a multivariate categorical distribution MC ( x c j | π ) parametrized by weights π = ( π c ) c C q and defined by
MC ( x c j | π ) = c c q π c δ x c j c
where c = ( c 0 , , c q 1 ) C q = C 0 × × C q 1 :
c C q π c = 1 and δ x c j c = 1 if x c j 0 = c 0 , , x c j q 1 = c q 1 0 otherwise .
This multivariate categorical distribution is proposed to tackle issues related to missing data by modelling a dependence structure for x c j that enables inference on missing categorical features.

2.3. Outlier Handling

Outliers are only considered for continuous data x q = ( x q j ) j = 1 J since only reliable categorical variables are assumed to be filled in databases and unreliable ones are processed as missing data. Then, continuous outliers are handled by introducing scale latent variables u = ( u j ) j = 1 J conditionally to categorical data x c due to the dependence structure established in (1) such that
j { 1 , , J } , x q j | u j , x c j c C q N x q j | μ c , u j 1 Σ δ x c j c and u j | x c j c C q G u j | α c , β c δ x c j c ,
where each u j follows conditionally to categorical data x c j a Gamma distribution with rate and shape parameters ( α c , β c ) R * + × R * + .

2.4. Missing Data Handling

Both continuous and categorical data ( x q j , x c j ) j = 1 J can be partially observed. Hence ( x q j , x c j ) j = 1 J are decomposed into observed features ( x q j obs , x c j obs ) j = 1 J and missing features ( x q j miss , x c j miss ) j = 1 J such that
j { 1 , , J } , x q j = x q j miss x q j obs with ( x q j miss , x q j obs ) R d j miss × R d j obs and d j miss + d j obs = d , x c j = x c j miss x c j obs with ( x c j miss , x c j obs ) C q j miss × C q j obs and q j miss + q j obs = q .
where ( R d j miss , C q j miss ) and ( R d j obs , C q j obs ) , are disjoint subsets of ( R d , C q ) embedding missing features ( x q j miss , x c j miss ) and observed features ( x q j obs , x c j obs ) . Missing continuous data x q miss = ( x q j miss ) j = 1 J are handled by taking advantage of properties of the multivariate normal distribution to obtain a distribution for missing values. Due to the dependence structure established in (1), missing continuous data x q miss = ( x q j miss ) j = 1 J are distributed conditionally to observed continuous data x q obs = ( x q j obs ) j = 1 J and categorical data x c as follows
j { 1 , , J } , x q j miss | x q j obs , x c j c C N x q j miss | μ j c x q miss , Σ x q miss δ x c j c , x q j obs | x q j obs , x c j c C N x q j obs | μ j c x q obs , Σ x q obs δ x c j c ,
where j { 1 , , J } , c C q :
μ j c x q miss = μ c miss + Σ cov Σ obs 1 x q j obs μ c obs , μ j c x q obs = μ c obs , Σ x q miss = Σ miss Σ cov Σ obs 1 Σ cov and Σ x q obs = Σ obs 1 + 2 × Σ obs 1 Σ cov Σ x q miss 1 Σ cov Σ obs 1 1 .
Noting that the dependence structure between categorical features is modeled through Kronecker symbols ( δ x c j c ) c C q , this dependence structure can be exploited to handle missing features such that the missing features x c j miss follow a multivariate categorical distribution conditionally to observed features x c j obs given by
p ( x c j miss = c miss | x c j obs = c obs ) = π c miss , c obs c miss c q j miss π c miss , c obs
where π c miss , c obs is the joint probability π c defined in (2) for c = ( c miss , c obs ) c q j miss × c q j obs .

3. Model and Inference

In this section, the proposed model is briefly presented as a hierarchical latent variable model handling missing values and outliers. Then, the inference procedure is developed through a variational Bayesian approximation. At last, classification and clustering algorithms are introduced by using the proposed model.

3.1. Model

According to a dataset x obs of i.i.d observations, independent latent variables h = ( x miss , u , z ), parameters Θ = ( a , π , α , β , μ , Σ ) of the K clusters and assumptions on mixed data defined in Section 2.1, the complete likelihood of the proposed mixture model can be expressed as
p ( x obs , h | Θ , K ) = j = 1 J k = 1 K a k c C q π k c N x q j miss x q j obs | μ k c , u j 1 Σ k G ( u j | α k c , β k c ) δ x c j miss , x c j obs c δ z j k
where
  • x obs = ( x q j obs , x c j obs ) j = 1 J are the observed features,
  • x miss = ( x q j miss , x c j miss ) j = 1 J are the latent variables modelling the missing features,
  • z = ( z j ) j = 1 J the independent labels for continuous and categorical observations x = ( x q j , x c j ) j = 1 J
  • u = ( u j ) j = 1 J the scale latent variables handling outliers for quantitative data x q and distributed according to a Gamma distribution with shape and rate parameters ( α , β ) = ( α k c , β k c ) ( k , c ) { 1 , , K } × C q ,
  • a = ( a k ) k = 1 K are the weights related to component distributions,
  • ( μ , Σ ) = ( ( μ k c ) c C q , Σ k ) k = 1 K the mean and the variance parameters of quantitative data x q for each cluster,
  • π = ( π k ) k = 1 K the weights of the multivariate Categorical distribution of categorical data x c for each cluster.
Eventually, the Bayesian framework imposes to specify a prior distribution p ( Θ | K ) for Θ which is chosen as
p ( Θ | K ) = p ( a | K ) p ( π | K ) p ( α , β | K ) p ( μ , Σ | K ) = D ( a | κ 0 ) k = 1 K D ( π k | π 0 ) c C q p ( α k c , β k c | p 0 , q 0 , s 0 , r 0 ) N μ k c | μ 0 k c , η 0 k c 1 Σ k IW Σ k | γ 0 , Σ 0
where D ( · | · ) and IW ( · | · ) denote the Dirichlet and Inverse-Wishart distributions and p ( · , · | p , q , s , r ) is a particular distribution designed to avoid a non-closed-form posterior distribution for ( α , β ) such that ( α , β ) R * + × R * + , p ( α , β | p , q , s , r ) p α 1 e q β β s α Γ ( α ) r .

3.2. Variational Bayesian Inference

The intractable posterior distribution P = p ( h , Θ | x obs , K ) is approximated by a tractable one Q = q ( h , Θ | K ) whose parameters are chosen via a variational principle to minimize the Kullback-Leibler (KL) divergence
K L Q | | P = q ( h , Θ | K ) log q ( h , Θ | K ) p ( h , Θ | x obs , K ) h Θ = log p ( x obs | K ) L ( q | K )
with L ( q | K ) a lower bound for the log evidence log p ( x obs | K ) given by
L ( q | K ) = E h , Θ log p ( x obs , h , Θ | K ) E h , Θ log q ( h , Θ | K ) ,
where E h , Θ [ · ] denotes the expectation with respect to q ( h , Θ | K ) . Then, minimizing the KL divergence is equivalent to maximizing L ( q | K ) . Assuming that q ( h , Θ | K ) can be factorized over the latent variables h and the parameters Θ , a free-form maximization with respect to q ( h | K ) and q ( Θ | K ) leads to the following update rules :
VBE step : q ( h | K ) exp E Θ log p ( x obs , h | Θ , K ) , VBM step : q ( Θ | K ) exp E h log p ( x obs , h , Θ | K ) .
Thereafter, the algorithm iteratively updates the variational posteriors by increasing the bound L ( q | K ) . Even if latent variables h and parameters Θ are assumed to be independent a posteriori, their conditional structures are preserved as follows
q ( h | K ) = q ( x q miss | u , x c miss , z , K ) q ( u | x c miss , z , K ) q ( x c miss | z , K ) q ( z | K ) , q ( Θ | K ) = q ( a | K ) q ( π | K ) q ( α , β | K ) q ( μ , Σ | K ) .
Eventually, the following conjugate variational posterior distributions are obtained according to the previous assumptions
q ( h | K ) = j = 1 J k = 1 K r ˜ j k c miss C q j miss r ˜ j k c miss x c miss c obs C q j obs N x q j miss | μ ˜ j k c x q miss , u j 1 Σ ˜ k x q miss G u j | α ˜ j k c , β ˜ j k c δ x c j obs c obs δ x c j miss c miss δ z j k , q ( Θ | K ) = D ( a | κ ˜ ) k = 1 K D ( π | π ˜ k ) c C q p ( α k c , β k c | p ˜ k , q ˜ k , s ˜ k , r ˜ k ) N μ k c | μ ˜ k c , η ˜ k c 1 Σ k IW ( Σ k | γ ˜ k , Σ ˜ k ) .
Their respective parameters are estimated during the VBE and VBM-steps by developing expectations E Θ log p ( x obs , h | Θ , K ) and E h log p ( x obs , h , Θ | K ) .

3.3. Classification and Clustering

According to the degree of supervision, three problems can be distinguished: supervised classification, semi-supervised classification and unsupervised classification known as clustering. The supervised classification problem is decomposed into a training step and a prediction step. The training step consists in estimating parameters Θ given the number of classes K and a set of training data x with known labels z . Then, the prediction step results in associating label z * of a new sample x * to its class k * chosen as the Maximum A Posteriori (MAP) solution
k * = arg max k = 1 K q ( z * = k | K )
given the previous estimated parameters Θ . In the semi-supervised classification, only the number of classes K is known and both labels z of the dataset x and parameters Θ have to be determined. As for the prediction step, the MAP criterion is retained for affecting observations to classes such that
k * = arg max k = 1 K q ( z = k | K ) .
Given a set of data x , the clustering problem aims to determine the number of clusters K ˜ , labels z of data and parameters Θ . Selecting the appropriate K ˜ seems like a model selection issue and is usually based on a maximized likelihood criterion given by
K ˜ = arg max K log p ( x | K ) = arg max K log p ( x , Θ | K ) d Θ .
Unfortunately, log p ( x | K ) is intractable and the lower bound in (3) is preferred to penalized likelihood criteria [8,15,16] since it does not depend on asymptotical assumptions and does not require Maximum Likelihood estimates. Then according to an a priori range of numbers of clusters { K min , , K max } , the semi-supervised classification is performed for each K { K min , , K max } and both z K and Θ K are estimated. Finally, the number of classes K ˜ in (4) is chosen as the maximizer of the lower bound L ( q | K ) :
K ˜ = arg max K L ( q | K ) .
After determining K ˜ , only z K ˜ and Θ K ˜ are kept as estimated labels and parameters.

4. Application

In this section, the proposed method is performed on a radar emitter dataset. For comparison, a standard neural network (NN), the k-nearest neighbours (KNN) algorithm, Random Forests (RdF) the k-means algorithm are also evaluated. Two experiments are carried out to evaluate classification and clustering performance with respect to a range of percentages of missing values.

4.1. Data

Realistic data are generated from an operational database gathering 55 radar emitters presenting various patterns. Each pattern consists of a sequence of pulses which are defined by a triplet of continuous features (pulse features) and a quartet of categorical features (pulse modulations) listed among 42 combinations of the categorical features. For each radar emitter, 100 observations ( x j ) j = 1 100 are simulated from its pattern of pulses such that an observation x j = ( x q j , x c j ) is made up of continuous features x q j and categorical features x c j related to one of the pulses. Extra missing values are added to evaluate limits of the proposed approach by randomly deleting coordinates of ( x q j ) j = 1 100 and ( x c j ) j = 1 100 for each of the 55 radar emitters. Therefore, imputation methods [17] are used to handle missing data for comparison algorithms. As for continuous missing data, they are handled through the Mean and k-nearest neighbours imputation methods whereas missing categorical data are handled through the k-nearest neighbours and mode imputation methods. These imputation methods are compared with the proposed approach where missing continuous data are reconstructed through the variational posterior marginal mean of missing continuous data given by j { 1 , , J } ,
x ˜ q j miss = E x qj miss q ( x q j miss , u j , x c j z j ) u j x c j z j = k = 1 K r ˜ j k c obs C q j obs δ x c j obs c obs c miss C q j miss r ˜ j k c miss x c miss μ ˜ j k c obs c miss x q miss
and missing categorical data are reconstructed through the variational posterior marginal mode of missing categorical data given by j { 1 , , J } ,
x ˜ c j miss = arg max c miss C q j miss q ( x c j miss , z j ) d z j = arg max c miss C q j miss k = 1 K r ˜ j k r ˜ j k c miss x c miss .

4.2. Classification Experiment

The classification experiment evaluates the ability of each algorithm to assign unlabeled data to one of the K classes trained by a set of labeled data. Since comparison algorithms do not handle datasets including missing values, a complete dataset is used to enable their training. During the prediction step, incomplete observations are completed thanks to the mean and KNN imputation methods and the posterior reconstructions defined in (6) and (7). For the classification experiment, results are shown in Figure 1. Without missing data, both algorithms cannot perfectly classify the 55 radar emitters for the 2 datasets. Indeed, both algorithms reach accuracies of 90 % for the continuous dataset and 98 % for the mixed dataset. These performance can be explained by the non total separability of continuous and categorical datasets since the 55 emitters share 42 combinations of categorical features and some intervals of continuous features. Nonetheless when mixed data are taken into consideration, the dataset becomes more separable leading to higher performance of both algorithms. When the proportion of missing values increases, the proposed model outperforms comparisons algorithms for each dataset. It achieves accuracies of 80 % and 95 % for 90 % of deleted continuous and mixed values whereas accuracies of comparison algorithms are lower than 65 % and 75 % with missing data imputation from standard methods. These higher performance of the proposed model reveal that the proposed method embeds a more efficient inference method than other imputation methods. That result is confirmed on Figure 1 when comparison algorithms are applied on data reconstructed by the proposed model. Indeed when the proposed inference is chosen, comparison algorithms share the same performance than the proposed model and manage to handle missing data even for 90 % of deleted values.
Then, effectiveness of the proposed model can be explained by the fact that missing data imputation methods can create outliers that deteriorate performance of classification algorithms whereas the inference on missing data and labels prediction are jointly estimated in the proposed model. Indeed, embedding the inference procedure into the model framework allows properties of the model, such as outliers handling, to counterbalance drawbacks of imputation methods such as outlier creation.

4.3. Clustering Experiment

The clustering experiment tests the ability of each algorithm to find the true number of clusters K ˜ among { 35 , , 85 } . The lower bound (3) and the average Silhouette score [18] are criteria used to select the optimal number of clusters for the proposed model and the k-means algorithm. Results of the clustering experiment are visible on Figure 2 which presents numbers of clusters selected by the lower bound and average Silhouette scores for the proposed model and k-means algorithm according to different proportions of missing values and imputation methods. Without missing data, the correct number of clusters (K = 55) is selected by the two criteria for the k-means algorithm and the proposed model when continuous and mixed data are clustered. In presence of missing values, the average Silhouette score mainly selects K = 65 when the k-means algorithm is run on the 2 datasets completed by standard imputation methods. When, the k-means algorithm performs clustering on the posterior reconstructions, the average Silhouette score correctly selects K = 55 until 60 % of missing values for continuous data and 40 % of missing values for mixed data. Eventually when the proposed model does clustering, the two criteria select the correct number of clusters K = 55 until 70 % of missing values for continuous and mixed data. These results show two main advantages of the proposed model. As previously, the proposed model provides a more robust inference on missing data since the average Silhouette score chooses more representative number of clusters when the k-means algorithm is run on the posterior reconstructions than on data completed by standard imputation methods. Furthermore, since the lower bound criterion also selects the correct number of clusters as the average Silhouette score, it can be used as a valid criterion for selecting the optimal number of clusters and does not require extra computational costs as the Silhouette score since it is computed during the model parameter estimation. Finally, the proposed approach provides a more robust inference on missing data and a criterion for selecting the optimal number of clusters without extra computations.

5. Conclusions

In this paper, a mixture model handling both continuous data and categorical data is developed. More precisely, an approach based on the conditional Gaussian mixture model is investigated by establishing conditional relations between continuous and categorical data. Benefiting from a dependence structure designed for mixed-type data, the proposed model shows its efficiency for inferring on missing data, performing classification and clustering tasks and selecting the correct number of clusters. Since the posterior distribution is intractable, model learning is processed through a variational Bayesian approximation where variational posterior distributions are proposed for continuous and categorical missing data. Experiments point out that the proposed approach can handle mixed-type data even in presence of missing values and can outperform standard algorithms in classification and clustering tasks. Indeed the main advantage of our approach is that it enables the counterbalance of imputation methods drawbacks by embedding the inference procedure into the model framework.

References

  1. Bishop, C.M. Pattern Recognition and Machine Learning (Information Science and Statistics); Springer: Berlin/Heidelberg, Germany, 2006. [Google Scholar]
  2. Hartigan, J.A.; Wong, M.A. Algorithm AS 136: A k-means clustering algorithm. J. R. Stat. Soc. Ser. C (Appl. Stat.) 1979, 28, 100–108. [Google Scholar] [CrossRef]
  3. Ester, M.; Kriegel, H.P.; Sander, J.; Xu, X. A Density-Based Algorithm for Discovering Clusters in Large Spatial Databases with Noise; Kdd: Portland, OR, USA, 1996; Volume 96, pp. 226–231. [Google Scholar]
  4. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  5. Sander, J.; Ester, M.; Kriegel, H.P.; Xu, X. Density-Based clustering in spatial databases: The algorithm GDBSCAN and its applications. Data Min. Knowl. Discov. 1998, 2, 169–194. [Google Scholar] [CrossRef]
  6. Jain, A.K. Data clustering: 50 years beyond K-means. Pattern Recognit. Lett. 2010, 31, 651–666. [Google Scholar] [CrossRef]
  7. Troyanskaya, O.; Cantor, M.; Sherlock, G.; Brown, P.; Hastie, T.; Tibshirani, R.; Botstein, D.; Altman, R.B. Missing value estimation methods for DNA microarrays. Bioinformatics 2001, 17, 520–525. [Google Scholar] [CrossRef] [PubMed]
  8. Biernacki, C.; Celeux, G.; Govaert, G. Assessing a mixture model for clustering with the integrated completed likelihood. IEEE Trans. Pattern Anal. Mach. Intell. 2000, 22, 719–725. [Google Scholar] [CrossRef]
  9. Lawrence, C.J.; Krzanowski, W.J. Mixture separation for mixed-mode data. Stat. Comput. 1996, 6, 85–92. [Google Scholar] [CrossRef]
  10. Everitt, B. A finite mixture model for the clustering of mixed-mode data. Stat. Probab. Lett. 1988, 6, 305–309. [Google Scholar] [CrossRef]
  11. Andrews, D.F.; Mallows, C.L. Scale Mixtures of Normal Distributions. J. R. Stat. Soc. Ser. B (Methodol.) 1974, 36, 99–102. [Google Scholar] [CrossRef]
  12. Waterhouse, S.; MacKay, D.; Robinson, T. Bayesian methods for mixtures of experts. In Advances in Neural Information Processing Systems; MIT Press: Cambridge, MA, USA, 1996; pp. 351–357. [Google Scholar]
  13. Schleher, D.C. Introduction to Electronic Warfare; Technical report; Eaton Corp., AIL Div.: Deer Park, NY, USA, 1986. [Google Scholar]
  14. Richards, M.A. Fundamentals of Radar Signal Processing; McGraw-Hill Education: New York, NY, USA, 2005. [Google Scholar]
  15. Akaike, H. Information theory and an extension of the maximum likelihood principle. In Selected Papers of Hirotugu Akaike; Springer: New York, NY, USA, 1998; pp. 199–213. [Google Scholar]
  16. Schwarz, G. Estimating the dimension of a model. Ann. Stat. 1978, 6, 461–464. [Google Scholar] [CrossRef]
  17. García-Laencina, P.J.; Sancho-Gómez, J.L.; Figueiras-Vidal, A.R. Pattern classification with missing data: A review. Neural Comput. Appl. 2010, 19, 263–282. [Google Scholar] [CrossRef]
  18. Kaufman, L.; Rousseeuw, P.J. Finding Groups in Data: An Introduction to Cluster Analysis; John Wiley & Sons: Hoboken, NJ, USA, 2009; Volume 344. [Google Scholar]
Figure 1. Classification performance are presented for the proposed model (PM) in blue, the NN in red, the RnF in green and the KNN in cyan. For each figure, solid lines represent accuracies with a posteriori reconstructed missing data, doted dashed lines stand for accuracies with mean/mode imputation whereas dashed lines show accuracies with KNN imputation for the comparison algorithms.
Figure 1. Classification performance are presented for the proposed model (PM) in blue, the NN in red, the RnF in green and the KNN in cyan. For each figure, solid lines represent accuracies with a posteriori reconstructed missing data, doted dashed lines stand for accuracies with mean/mode imputation whereas dashed lines show accuracies with KNN imputation for the comparison algorithms.
Proceedings 33 00023 g001
Figure 2. Estimation of the number of clusters using the lower bound (LB) and the silhouette score (S) for the proposed model and only the silhouette score (S) for the k-means algorithm.
Figure 2. Estimation of the number of clusters using the lower bound (LB) and the silhouette score (S) for the proposed model and only the silhouette score (S) for the k-means algorithm.
Proceedings 33 00023 g002

Share and Cite

MDPI and ACS Style

Revillon, G.; Mohammad-Djafari, A. A Complete Classification and Clustering Model to Account for Continuous and Categorical Data in Presence of Missing Values and Outliers . Proceedings 2019, 33, 23. https://doi.org/10.3390/proceedings2019033023

AMA Style

Revillon G, Mohammad-Djafari A. A Complete Classification and Clustering Model to Account for Continuous and Categorical Data in Presence of Missing Values and Outliers . Proceedings. 2019; 33(1):23. https://doi.org/10.3390/proceedings2019033023

Chicago/Turabian Style

Revillon, Guillaume, and Ali Mohammad-Djafari. 2019. "A Complete Classification and Clustering Model to Account for Continuous and Categorical Data in Presence of Missing Values and Outliers " Proceedings 33, no. 1: 23. https://doi.org/10.3390/proceedings2019033023

Article Metrics

Back to TopTop