Next Article in Journal
An Industrial Load Classification Method Based on a Two-Stage Feature Selection Strategy and an Improved MPA-KELM Classifier: A Chinese Cement Plant Case
Next Article in Special Issue
Stylometric Fake News Detection Based on Natural Language Processing Using Named Entity Recognition: In-Domain and Cross-Domain Analysis
Previous Article in Journal
Cascading and Ensemble Techniques in Deep Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Supervised Dimensionality Reduction of Proportional Data Using Exponential Family Distributions

by
Walid Masoudimansour
1,* and
Nizar Bouguila
2
1
Concordia Institute for Information Systems Engineering (CIISE), Concordia University, Montreal, QC H3G 1S6, Canada
2
Department of Electrical and Computer Engineering, Concordia University, Montreal, QC H3G 1S6, Canada
*
Author to whom correspondence should be addressed.
Electronics 2023, 12(15), 3355; https://doi.org/10.3390/electronics12153355
Submission received: 25 June 2023 / Revised: 25 July 2023 / Accepted: 1 August 2023 / Published: 5 August 2023
(This article belongs to the Special Issue Data Push and Data Mining in the Age of Artificial Intelligence)

Abstract

:
Most well-known supervised dimensionality reduction algorithms suffer from the curse of dimensionality while handling high-dimensional sparse data due to ill-conditioned second-order statistics matrices. They also do not deal with multi-modal data properly since they construct neighborhood graphs that do not discriminate between multi-modal classes of data and single-modal ones. In this paper, a novel method that mitigates the above problems is proposed. In this method, assuming the data is from two classes, they are projected into the low-dimensional space in the first step which removes sparsity from the data and reduces the time complexity of any operation drastically afterwards. These projected data are modeled using a mixture of exponential family distributions for each class, allowing the modeling of multi-modal data. A measure for the similarity between the two projected classes is used as an objective function for constructing an optimization problem, which is then solved using a heuristic search algorithm to find the best separating projection. The conducted experiments show that the proposed method outperforms the rest of the compared algorithms and provides a robust effective solution to the problem of dimensionality reduction even in the presence of multi-modal and sparse data.

1. Introduction

Lately, due to a significant decrease in the cost of deploying sensors and other data collection devices, there has been a surge in the way these devices are used for collecting samples of data. Consequently, there is tremendous redundancy in the stored data and this has been a source of problem for many applications such as data analysis, classification, and clustering. These problems include, but are not limited to, the difficulty of visualization, introducing the curse of dimensionality [1,2], and an excessive need for processing and storage resources. Feature selection techniques [3,4] are a group of methods that are used to remedy this problem; however, dimensionality reduction (DR) methods prove to be a more efficient way of reducing this redundancy while preserving most of the embedded information [5,6,7] since they use a combination of features instead of a subset of them. Such techniques are well studied and researched and are diverse in the approach they adopt to solve the problem. From one general point of view, these algorithms are divided into linear and non-linear types. In case of linear algorithms, a linear transform is used to project the data to the low-dimensional space. These projections are the most popular due to the simplicity of the final transform and the intuitive interpretation of the projection as a weighted sum of the original features and often are based on the second-order statistics of the data [8,9,10,11]. The second-order statistics of data refer to statistical properties that are derived from the covariance matrix of the data, and involve analyzing the relationships between different data features and how they vary together. The covariance matrix captures the interdependencies between variables, providing valuable information about the data’s distribution and structure. In the context of the DR methods, the use of second-order statistics enables an algorithm to preserve important discriminative information and enhance the separability of classes in the reduced feature space. Fisher Discriminant Analysis (FDA) [12], Principal Component Analysis (PCA) [13], Local Fisher Discriminant Analysis (LFDA) [14,15], Factor Analysis (FA) [16], Locality Preserving Projection (LPP) [17], and L1-Norm Fisher Discriminant Analysis (LDA-L1) [18] are some of the well-studied ones among them. In the non-linear category of DR techniques falls any other method that does not use a linear projection for mapping the data into the low-dimensional space. For example, Kernel PCA (KPCA) [19], Maximum Variance Unfolding (MVU) [20], Uniform Manifold Approximation and Projection (UMAP) [21], t-distributed stochastic neighbor embedding (t-SNE) [22], Locally Linear Embedding (LLE) [23], and Kernel-based Within-Class Collaborative Preserving Discriminant Projection (KWCCPDP) [24] are some of the most well known among such methods. These techniques might be a better match to the problem depending on the type of the data and their non-linear nature. From another point of view, DR algorithms can be divided into supervised/unsupervised depending on whether they use available data labels for projecting the data into the low-dimensional space. Of course, the usage of data labels allows supervised DR techniques to capture the hidden structures in the data more effectively.
In this research, we limit our focus to the linear supervised techniques since they are highly efficient and capture the structures in the data very well. However, most of such methods suffer from some weaknesses. For example, some may not be able to handle multi-modal data effectively (for such analysis regarding LPP, see [14]). Also, some (e.g., LFDA and SOLPP) completely fail to produce any results in certain cases due to the sparsity of the data as has been shown in Section 4. While dealing with extremely sparse data, another important issue with many of the conventional algorithms is that the effectiveness of the solution highly depends on the choice of the training data. This is due to the fact that they depend on solving a generalized eigenvalue problem (GEVP), which is more likely to become ill posed when the data are highly sparse.
The above-mentioned weaknesses while dealing with sparse high-dimensional data motivated the design of a DR technique that uses a different approach to tackle the dimensionality reduction problem. This novel DR technique is designed to work best for proportional data with very high dimensions, and specifically designed for classification problems. Its goal is to maximize the separation of data in the new low-dimensional space. Starting with the projection of data classes into the low-dimensional space, it assumes each class of data in this new space has a distribution that is modeled with a mixture of distributions from the exponential family. This mixture allows multi-modal data to be modeled and is one of the advantages of this algorithm. After estimating the parameters of this mixture for each class using the EM algorithm, we use an approximation of the KL (Kullback–Leibler) divergence between the mixtures to estimate their distance. This distance is then used as a measure to be maximized by the algorithm. To this end, one can use an optimization technique to maximize this measure and hence optimize data separation; however, the use of the EM algorithm limits the methods of optimization that can be used efficiently. Specifically, heuristic search methods remain the best option in this case, and we have used GA (genetic algorithm) effectively to solve the problem and find a near-optimal projection for data separation. As mentioned above, the proposed technique uses distributions from the exponential family. The choice of the distribution depends on the type of the data and the hidden structures, e.g., the structure of the covariance matrix. Nevertheless, as has been shown in the literature, some distributions can model proportional data effectively [25,26,27,28,29,30,31] and we have used such distributions. It is important to note that any desirable subset of distributions from the set of exponential family distributions can be used in the proposed method. In general, our contributions is this paper can be summarized as follows:
  • Proposing a novel supervised dimensionality reduction method that addresses the curse of dimensionality in high-dimensional sparse proportional data by projecting data into a low-dimensional space, mitigating sparsity significantly.
  • Introducing a unique approach to handle multi-modal data by modeling the projected data using a mixture of exponential family distributions for each class, allowing for effective discrimination between multi-modal and single-modal classes.
  • Formulating a closed form for the similarity between projected classes using KL-Divergence and employing a heuristic search algorithm to optimize the separation of classes, resulting in a robust and efficient solution to the problem of dimensionality reduction, outperforming other compared algorithms in diverse experimental settings.
The remainder of this work is structured as follows. In Section 2, some conventional techniques of DR are introduced which are used in the experiments for comparison purposes to show the effectiveness of the designed method. The problem is formulated in detail in Section 3. In Section 4, different datasets are used to test the performance of the proposed technique. Finally, we draw our conclusion and present some final words in Section 5.

2. Related Works

As has been mentioned before, conventional linear DR methods, despite all their differences, use the second-order statistics of data to formulate and solve a GEVP. In this section, we will briefly discuss some of the well-known supervised linear DR techniques that are used in the literature, and will use them to assess the performance of our algorithm later in Section 4.
Supervised Locality Preserving Projection (SLPP) [32] is an extension of the Locality Preserving Projection (LPP) method and is adapted to use the labels of the data. In LPP, first, a matrix A with elements a i , j [ 0 , 1 ] is defined to reflect the neighborhood of the data, and, based on this matrix, a transform T is defined as follows.
T = argmin T R N × K 1 2 i = 1 M j = 1 M a i , j T x i T x j 2 s . t . T X D X T = I
where M is the number of data points, N and K are the original and target dimensions respectively, and  x i denotes a data point. Moreover, the elements of D are
d i , j = j = 1 M a i , j if   i = j 0 if   i j
Solving the bellow GEVP, one can find the transform T.
X ( D A ) X v = λ X D X v
By choosing the K eigenvectors corresponding to the K smallest eigenvalues of the above GEVP, the transform T will reduce the dimensionality of the data to K.
LPP and its derivatives suffer from several problems, including overfitting, sensitivity to noise, and disconnectivity of the neighborhood graph. In [33], the authors introduce Supervised Optimal Locality Preserving Projection (SOLPP) to tackle these problems. This has been accomplished by defining a discriminating similarity matrix between data points called W. The following constrained optimization problem is then formulated, where i j and t i indicates the i-th row of T.
minimize T trace ( T X ( D W ) X T ) s . t . T T = I , t i X D X t j = 0 , t j X X t i = 0
By solving this problem using an iterative method, the optimal projection T can be evaluated.
Locality Sensitive Discriminant Analysis (LSDA) [34] is based on two connectivity graphs, corresponding to inter-class and intra-class distances. By incurring heavy penalty for neighboring points from the same class being mapped far from each other, or nearby points from different classes being mapped close to each other, the authors formulate an optimization problem and find an optimal projection.
Fisher Discriminant Analysis (FDA) starts by formulating the inter-class and intra-class scatter matrices as follows:
S W = 1 2 i , j = 1 M a i , j ( W ) ( x i x j ) ( x i x j )
S B = 1 2 i , j = 1 M a i , j ( B ) ( x i x j ) ( x i x j )
where
a i , j ( W ) = 1 m l if   l i = l j = l 0 if   l i l j
a i , j ( B ) = 1 M 1 m l if   l i = l j = l 1 M if   l i l j
Here, l i and m l are the label of the sample x i and the number of samples in class l, respectively. Solving the following optimization problem, one can arrive at the optimal projection of FDA.
T = argmax T R N × K t r ( T S W T ) 1 ( T S B T )
The optimal transform T is actually formed by the K eigenvectors corresponding to the K smallest eigenvalues of the following GEVP.
S B v = λ S B v
The authors in [14], on the other hand, use an affinity matrix similar to LPP to rebuild the scatter matrices a ¯ ( W ) and a ¯ ( B ) as below.
a ¯ i , j ( W ) = a i , j 1 m l if   l i = l j = l 0 if   l i l j
a ¯ i , j ( B ) = a i , j ( 1 M 1 m l ) if   l i = l j = l 1 M if   l i l j
Replacing these matrices with those of Equations (6) and (7), a new GEVP can be solved to find the optimal projection corresponding to LFDA.
Finally, in [18], the authors introduce a new variant of LDA by replacing the L2 norm with the L1 norm. This will make the algorithm more robust to outliers and sparsity; however, the resulting problem is non-convex and therefore an iterative method must be used to find the optimal projection.
All the above-mentioned methods of DR have been proved to be efficient in the literature. Nevertheless, they fail to effectively solve the problem of dimensionality reduction in some cases. For instance, they mostly rely on the second-order statistics of data and solving a GEVP. For very high-dimensional data (and specifically for the case of proportional data), however, the sparsity of the data matrix results in near-singular correlation matrices that cause the GEVP to be ill posed and not result in an optimal solution. Moreover, the above methods majorly rely on the construction of a neighborhood graph that sometimes becomes disconnected and affects the solution of the GEVP. Multi-modal data representation is another important task in which the above methods often fail [14]. Note that in all of the above methods, the algorithm has to manipulate very large-size matrices, and, for example, invert large sparse matrices, which is inefficient and of the order of O ( N 3 ) in terms of complexity. Therefore, the complexity of the problem can grow rapidly by the number of features. Finally, note that none of the mentioned algorithms are designed for proportional data specifically and therefore the projected data may not respect the proportionality condition.
The above-mentioned problems with the current supervised linear DR methods motivate us to introduce EXMMP. This algorithm is designed specifically as a pre-processing step in classification problems. Therefore, easy separation of data classes is of utter importance in this algorithm. Since we start with mapping the data using a projection matrix to the low-dimensional space, at this very first step, the problem of sparsity is resolved to a good extent. The complexity of the algorithm, on the other hand, is not affected heavily by the number of original features or the number of available samples. Note that the heuristic search for the optimal projection embeds a property for the projection that guarantees the proportionality of the mapped data, and therefore, unlike the conventional methods, maps proportional data to a corpus of proportional data. As the final advantage, the usage of mixture models allows the algorithm to model multi-modal data, where the rest of the algorithms are unable to do so and fail to capture this property.

3. Proposed Method

3.1. Problem Statement

In this section, we will state the problem and formulate our method, which resolves the issue of the sparsity of the data that hinders other algorithms in reducing the dimensionality of data properly. We use a mixture of distributions from the exponential family along with an estimation of the KL-Divergence between two mixtures—a similarity measure to optimize the separability of classes.
Consider a set of proportional data containing M samples of N dimensions. Let us represent each sample with x i where
x i , j 0 , j = 1 N x i , j = 1
and 1 i M , 1 j N . Also, x i , j is the j-th element of the i-th sample x i . To facilitate the manipulation of this data set, the matrix X is formed by accumulating x i s as its columns. The proposed algorithm maps this corpus of data from the N-dimensional space to a smaller K-dimensional space where K < N . Representing this projection with P, the new K-dimensional data set Y can be found as follows.
Y = P X
Note that after the projection, columns of Y represent the corresponding data samples in the original data matrix. Denote the elements of P by p r , s , where 1 r K and 1 s N . Moreover, to maintain the proportionality of the projected samples, we assume the following property for P (which is both necessary and sufficient for the projected data to be proportional [30]):
p r , s 0 , r = 1 K p r , s = 1 , 1 r K , 1 s N

3.2. Class Distribution Estimation

In this step, we assume that the projected data of each class are generated using a specific distribution. This distribution will then be used, along with its counterpart from the other class, to maximize data separation in the low-dimensional space. Furthermore, we are going to make the assumption that this generating distribution is one of the distributions from the exponential family and has a support on a simplex. These two assumptions are crucial since the former facilitates the derivation of the results and the latter conforms to the fact that the data are proportional [25]. Note that henceforward, it is assumed that each class of (projected) data is modeled with a mixture of a specific distribution. This assumption is also critical for the capacity of the algorithm for modeling multi-modal data. In the following, the EM (Expectation–Maximization) algorithm is used for estimating the parameters of such mixture model.
Representing the projected samples by y k = ( y 1 , k , y 2 , k , , y K , k ) T , their distribution can be written as
p ( y k | Θ ) = j = 1 Q ϕ j p j ( y k | θ j )
where i = 1 K y i , k = 1 , 0 y i , k 1 ; the vertical vector θ j contains the corresponding parameters of the j-th component; the matrix Θ , of which the columns are θ j s, represents the complete set of the parameters of the mixture; and the number of components in the mixture is represented by Q. Also, note that the prior ϕ j s conform to the following:
j = 1 Q ϕ j = 1
Now, assuming two training sample sets C 0 and C 1 are chosen from two classes of data such that C 0 = M 0 and C 1 = M 1 , the likelihood of the data given the parameter Θ ( κ ) , κ { 0 , 1 } is as follows
L ( Y ( κ ) | Θ ( κ ) ) = i = 1 M κ j = 1 Q ϕ j ( κ ) p j ( y i ( κ ) | θ j ( κ ) )
where samples are assumed to be i.i.d., y i ( κ ) is the i-th sample from class κ and the matrix Y ( κ ) is constructed by a column-wise population of all the class samples. Now, using the EM algorithm and latent variables Z i , j , which are indices for mixture assignment, one can find Θ ( κ ) such that this likelihood is maximized. First, the mixing weights are estimated from the equation
ϕ j ( κ ) ( t + 1 ) = i = 1 M κ Z ^ i , j ( κ ) ( t ) M κ
where t is the iteration number and Z ^ i , j ( κ ) ( t ) is the expected value of Z i , j ( κ ) ( t ) .
In the next step, to maximize the log likelihood, θ j ( κ ) ( t + 1 ) must be evaluated by solving the below set of equations
i = 1 M κ Z ^ i , j ( κ ) ( t ) θ j , k ln ( p j ( y i ( κ ) | θ j ) ) = 0
where 1 k N P , N P is the number of the parameters of the distribution, 1 j Q , and  θ j , k is the k-th element of θ j .
Let us assume the following distribution from the exponential family:
p ( y k | θ j ) = h ( y k ) e η j T · T ( y k ) A ( η j )
where the base measure h ( · ) (a real-valued function) and sufficient statistics T ( · ) (a vector-valued functions) are independent of the distribution parameter, η j (the natural parameter) is a vector-valued function of θ j , and A ( η j ) (the log partition) is a real-valued function of η j . Using this distribution in the EM algorithm, the maximization step yields the following system of equations for 1 j Q and 1 k N P :
i = 1 M κ Z ^ i , j ( t ) η j , k log h ( y i ) e η j T · T ( y i ) A ( η j ) = 0
where η j , k indicates the k-th element of η j and N P is the number of elements in the natural parameter. Simplifying the above system results in a set of N P × Q equations
A ( η j ) η j , k = i = 1 M κ Z ^ i , j T k ( y i ) i = 1 M κ Z ^ i , j
where T k ( · ) is the k-th element of the sufficient statistics function T ( · ) . Due to the non-linearity of the above system, no closed-form solution is available to find the maximizing parameters. Therefore, some type of numerical method is required to find the result. In this paper, the Newton–Raphson (NR) algorithm, which is one of the most well-known methods, has been used to find the parameters of the distribution. This method uses the Jacobian of the system which is the Hessian matrix of the original system, and hence will be called H for 1 k , s N P and 1 j Q . This matrix only depends on the log partition function and is evaluated below.
H = 2 A ( η j ) η j , k η j , s
Note that since A ( η a ) is independent of η b for a b , the resulting Jacobian matrix will be block diagonal which is a significant advantage since inverting this matrix will be of order of complexity O ( N P 3 ) instead of O ( Q 3 N P 3 ) (assuming cubic complexity for inverting a matrix).

3.3. Measuring Inter-Class Distance

To facilitate separation of the data, we develop a measure of distance between the two projected classes. One of the most well-known measures of distance between distributions is the KL-Divergence [35], and it has been used in this section. Finding a closed form for evaluating the KL-Divergence between two mixtures of distributions is not trivial. To do so, we follow a method similar to the one in [36]. Assuming two mixtures f ( y ) and g ( y ) consisting of Q f and Q g components, respectively, we start with the following equations.
f ( y ) = i = 1 Q f w i f i ( y ) , i = 1 Q f w i = 1 g ( y ) = j = 1 Q g u j g j ( y ) , j = 1 Q g u j = 1
Based on the definition of KL-Divergence, one arrives at
KL ( f ( y ) , g ( y ) ) = f ( y ) ln f ( y ) g ( y ) d y = f ( y ) ln f ( y ) d y f ( y ) ln g ( y ) d y
The first term of the above equation can be expanded as follows.
f ( y ) ln f ( y ) d y = i = 1 Q f w i f i ( y ) ln j = 1 Q f w j f j ( y ) d y = i = 1 Q f w i f i ( y ) ln j = 1 Q f w j f j ( y ) d y
A maximum lower bound for the left side of the above statement can be found using Jensen’s inequality, adjusting the introduced parameters ζ i , j in the equation below.
i = 1 Q f w i f i ( y ) ln j = 1 Q f w j f j ( y ) d y = i = 1 Q f w i f i ( y ) ln j = 1 Q f ζ i , j w j f j ( y ) ζ i , j d y i = 1 Q f w i f i ( y ) j = 1 Q f ζ i , j ln w j f j ( y ) ζ i , j d y
Note that for the above to be true, the following condition must hold.
j = 1 Q f ζ i , j = 1 , i { 1 , 2 , , Q f }
To find the proper ζ i , j parameters, we form the following cost function:
J ¯ = i = 1 Q f w i f i ( y ) j = 1 Q f ζ i , j ln w j f j ( y ) ζ i , j d y + i = 1 Q f λ i j = 1 Q f ζ i , j 1 + i = 1 Q f μ i ( ζ i , j )
for which the KKT (Karush–Kuhn–Tucker) conditions [37] are as below.
J ¯ ζ i , j = 0 , j = 1 Q f ζ i , j = 1 ζ i , j 0 , μ i 0 μ i ( ζ i , j ) = 0 , 1 i Q f
The solution to this optimization problem is
ζ i , j = w j e H ( f i , f j ) k = 1 Q f w k e H ( f i , f k )
where we define the function H ( · , · ) as
H ( f i , f j ) = f i ( y ) ln f j ( y ) d y
Replacing Equation (30) in the right-hand side of Equation (27), we will have
f ( y ) ln f ( y ) d y i = 1 Q f w i ln k = 1 Q f w k e H ( f i , f k )
Similarly, a maximized lower bound can be found for the second term of Equation (25), which is as below.
f ( y ) ln g ( y ) d y i = 1 Q f w i ln k = 1 Q g u k e H ( f i , g k )
Using the resulting two maximized lower bounds and noting that
KL ( f i , f j ) = H ( f i , f i ) H ( f i , f j )
a closed-form approximation can be found for the KL-Divergence of two mixtures of distributions.
KL ˜ ( f ( y ) , g ( y ) ) = i = 1 Q f w i ln k = 1 Q f w k e H ( f i , f k ) k = 1 Q g u k e H ( f i , g k ) = i = 1 Q f w i ln k = 1 Q f w k e KL ( f i , f k ) k = 1 Q g u k e KL ( f i , g k )
Note that this allows the KL-Divergence of two mixtures to be approximated in a closed form using the KL-Divergence of the two corresponding single distributions.
The KL-Divergence of two distributions f i and f j from the exponential family with parameters θ i and θ j , respectively, is as below.
KL ( f i , f j ) = E f i ln f i f j = E f i ln f i E f i ln f j
Note that E f i · denotes the expected value with respect to f i . To simplify the first term on the right-hand side of Equation (36), we plug Equation (20) and obtain
E f i ln f i = E f i ln h ( y ) E f i A ( θ i ) + k = 1 K η k ( θ i ) A ( η ) η k | θ i
Similarly, for the second term,
E f i ln f j = E f i ln h ( y ) E f i A ( θ j ) + k = 1 K η k ( θ j ) A ( η ) η k | θ i
And therefore, the closed form of the KL-Divergence will reduce to
KL ( f i , f j ) = A ( θ j ) A ( θ i ) + k = 1 K [ η k ( θ i ) η k ( θ j ) ] A ( η ) η k | θ i
After this final step, a closed-form approximation for the KL-Divergence of two mixtures of distributions can be obtained by using Equation (39) in Equation (35). Also, note that in this work we have used an approximation of the symmetric form of the KL-Divergence, namely, J-Divergence, which is formulated as follows.
J ˜ ( f i , f j ) = KL ˜ ( f i , f j ) + KL ˜ ( f j , f i ) = k = 1 K [ η k ( θ i ) η k ( θ j ) ] A ( η ) η k θ i A ( η ) η k | θ j ]

3.4. Maximizing Class Distance

To find the projection that maximizes the class distance between two projected classes, an optimization problem must be formulated. In the case of high-dimensional data, this results in solving for a significantly large number of parameters of the projection matrix, namely N × K . Considering that estimating the parameters of the two mixtures of distributions is performed using the EM algorithm, which is iterative, using conventional optimization techniques such as Gradient Descent (GD) is not practical in terms of complexity. Also, note that the fitness function has many local minima and, hence, the GD method is not suitable for this matter. Therefore, considering the nature of the problem, a heuristic search algorithm proves to be a proper method. In this paper, we use genetic algorithm (GA) along with the above-developed measure of distance as the fitness function. The initial population in the GA is a set of random matrices which conform to Equation (14). In each iteration, the data from both classes are projected to the low-dimensional space using one of the members of the current population, and the mixture parameters for each class of data are estimated and, then, used in Equation (40) to evaluate the fitness of the projecting matrix. This value is then used in the GA to find the optimum projection. Algorithm 1 briefly describes this process in the form of pseudo-code.
Algorithm 1 Summary of the proposed algorithm
1:
Input: Matrix of samples (X) and their labels
2:
Output: Optimum projection P
3:
Initialization:  P 1 , , P G G ( i ) = G , randomly initialized matrices that satisfy Equation (14)
4:
for generation G ( i )  do
5:
    for member j G ( i )  do
6:
        Find projected data Y j = P j X
7:
        Using EM and data from C κ , find Θ ( κ ) , j , κ { 0 , 1 }
8:
        Estimate the symmetric KL-Divergence between two estimated mixtures represented by Θ ( 0 ) , j and Θ ( 1 ) , j
9:
        Assign the above value as the fitness of P j
10:
    end for
11:
    Generate new population G ( i + 1 ) based on the evaluated fitness
12:
end for
13:
Chose the fittest member of the current generation as the optimal projection

4. Experimental Results

4.1. Test Scenarios

To prove the efficacy of the proposed technique, it is compared with several supervised linear DR methods in this section. As mentioned, EXPMMP uses distributions from the exponential family to model the data. However, distributions with support on the simplex are better suited for proportional/compositional data. In [38], Aitchison shows that an invertible function, specifically the log-ratio transformation, can be used to convert the support of any distribution to a simplex. For the experiments in this section, we used three distributions which are defined on a simplex: Dirichlet, Generalized Dirichlet, and Beta-Liouville. Moreover, we used the Additive Logistic Normal (ALN) distribution for our tests, which uses the above-mentioned log-ratio function along with the normal distribution. For all the performed experiments, a corpus of labeled data from two classes was mapped to a low-dimensional space using all the algorithms, followed by a Decision Tree (DT) classifier for classification. The choice of classes was made either randomly or based on the number of samples of that class in the dataset, where the classes with a high number of samples were chosen for reliable results. For the proposed method, we divided the data into training, validation, and testing sub-sets. This allowed us to perform the algorithm using different distributions and, after the training phase, choose the projection corresponding to the distribution that yielded the best classification accuracy on the validation sub-set. Note that the reported values in the tables of this section are all from test sub-set and each test was repeated 5 times while each repetition used a fivefold cross-validation method. The reported values in the tables are average accuracies and their corresponding standard deviation. Also, note that some algorithms are not present in the tables, which shows they have failed to produce any meaningful result.
Experiment 1: For the first experiment, the 20-Newsgroups dataset was used in the form of a bag-of-words model. This dataset contains 20,000 samples of text from 20 different classes and originally has 61,189 words. We chose two classes of data, namely atheism and forsale, and projected these classes into a low-dimensional space after reducing their features to 1000 by removing stop words and low-frequency features. Table 1 represents the resulting accuracies for the compared and proposed algorithms. Note that, the test was run also on SOLPP and LFDA, which did not yield any result since these algorithms could not solve the eigenvalue problem due to the sparsity of the data. Also, SLPP and LSDA reported very high deviation values in the tests, which is due to the fact that depending on the training sub-set, they may fail to solve the eigenvalue problem, again, because of the sparsity of the representing matrix. On the other hand, EXPMMP consistently reports a high accuracy of classification with a low value of standard deviation for different tests. To contextualize these results in relation to our objectives, we aimed to address the challenge of high-dimensional data classification while considering the sparsity inherent in the data. Our experiments demonstrate that EXPMMP successfully tackles this challenge, showcasing its potential as a promising solution for handling high-dimensional sparse data and achieving improved classification accuracy.
Experiment 2: To test the proposed algorithm for a different application, we used the Caltech-256 dataset, which consists of different samples of 256 objects. Some sample images from this dataset are presented in Figure 1. In this experiment, two image categories were chosen and their SIFT features were extracted. This set of features was used to construct a VBoW (Visual Bag of Words) using a dictionary of 750 words. The proposed method was then run on this binary classification problem along with the other compared algorithms with a target dimension of 3. Table 2 shows the results of this classification. As has been mentioned before, a DT was used to classify the reduced-dimension data and the mean and standard deviation of the resulting accuracies from all the tests have been reported. Note that the use of a low target dimension of 3 allows for a meaningful reduction in the feature space, effectively enhancing computational efficiency while preserving essential discriminative information. This aligns with our goal of developing an efficient yet accurate approach for high-dimensional data classification.
Note that the SOLPP and LFDA methods did not yield any results due to the sparsity of the data. Moreover, the proposed method performs better than the rest of the algorithms, consistently, in terms of both the average and standard deviation of accuracies, which shows that our proposed method is effective on diverse types of data.
Experiment 3: This experiment tested the proposed method using a different image dataset called Linnaeus-5 [39]. This dataset consists of images of five subjects (Berry, Bird, Dog, Flower, Other) each containing 1600 images. Following the same methodology used in the previous image classification experiment, we constructed a Visual Bag of Words (VBoW) representation with a dictionary size of 750 words. The extracted features were then employed for binary classification using the proposed method and several state-of-the-art algorithms as comparators. The results of this classification experiment are summarized in Table 3. Notably, the proposed method consistently outperforms the compared algorithms in terms of classification accuracy across all categories. This noteworthy improvement in accuracy highlights the effectiveness and robustness of our proposed approach, even when applied to a different and challenging image dataset like Linnaeus-5. By achieving superior accuracy consistently, our method showcases its potential for addressing real-world image classification tasks with diverse subject categories. The reliable performance of our algorithm on the Linnaeus-5 dataset reaffirms its generalizability and underscores its relevance in handling high-dimensional image data across various domains.
Experiment 4: In this experiment we, again, used the 20-Newsgroups dataset, although in a different setup. In this case, we fixed the target dimension to 3 while we tested different pairs of classes against each other. The preprocessing step left 1000 original features, which were reduced to 3 using all the algorithms. Table 4 presents the results of this pairwise classification test. As expected, the sparsity of the data posed challenges for certain algorithms. Specifically, SOLPP and LFDA encountered difficulties in finding a suitable low-dimensional representation for the data in this particular setup, leading to suboptimal performance or failure to provide meaningful results. In contrast, our proposed method consistently outperformed the compared algorithms in all cases, achieving superior classification accuracy for the pairwise tests. The ability of our method to handle the fixed target dimension of 3 while maintaining competitive performance underscores its adaptability and robustness in handling diverse experimental conditions. The outcomes of this experiment provide valuable insights into the strengths and limitations of the algorithms tested. The difficulties faced by SOLPP and LFDA emphasize the importance of choosing an appropriate method that aligns with the specific experimental setup and dataset characteristics (in this case, sparse compositional data). Moreover, the consistently better performance of our proposed method confirms its efficacy in capturing relevant discriminative information and effectively representing data in a low-dimensional space, even under challenging conditions.
Experiment 5: In this experiment, the ImageNet dataset was used to compare the proposed method with the rest of the algorithms. This dataset is composed of more than 20,000 categories with a different number of images in each category. Similar to Experiment 2, we chose two image categories in each test and constructed a VBoW with a 750-word dictionary. Then, we reduced the dimension of the data to 3 and used DT to classify the images. The results of the tests are presented in Table 5. Notably, we observed that our proposed method consistently achieved competitive results compared to the other algorithms across different category pairings. The ability of our method to maintain its performance on the vast and diverse ImageNet dataset underscores its potential as a robust and versatile approach for large-scale sparse image datasets. The ImageNet dataset poses significant challenges due to its diversity, with categories exhibiting varying image quantities and unique visual characteristics. Despite these complexities, our proposed method showed its capacity to capture meaningful discriminative information effectively, even when the representation of the images were highly sparse. Our objective for this experiment was twofold: firstly, to assess the performance of the proposed method on a widely recognized and complex image dataset, and secondly, to highlight its adaptability to highly sparse proportional data even when the target dimension is as low as 3.
Experiment 6: In this experiment, we used an image dataset known as Food-101 which contains images from 101 types of dishes. Similar to the previous experiments, we chose pairs of classes, constructed a VBoW for each pair using a 750-word vocabulary and SIFT features, and reduced the dimensionality to 3. The data were then classified and the test was repeated, and the mean and standard deviation of the accuracies were evaluated. Table 6 presents the results of this test. As observed, our proposed method consistently outperformed the compared algorithms across all class pairings. The ability of our method to maintain its superiority on the Food-101 dataset with its diverse visual characteristics underscores its effectiveness in image classification tasks under significantly sparse conditions. Notably, certain algorithms, such as SOLPP and LFDA, encountered challenges in producing meaningful results on this dataset. The difficulties faced by these algorithms highlight the importance of choosing appropriate methods that can handle the sparsity of the dataset. The consistently superior results achieved by our proposed method reaffirm its potential as a powerful tool for dimensionality reduction of proportional sparse data.
Example 7: In this test, we investigated the capacity of the introduced method for the case of multi-modal data. The previously used dataset, Food-101, was used in this case. After choosing three classes of data, two of them were combined as one multi-modal class and used for classification against the third. Similar to the previous tests, after constructing a VBoW model and reducing the dimensionality to 3, the samples were classified using a DT classifier. Table 7 presents the results of this multi-modal data test. Note that our proposed method consistently outperformed the compared algorithms in terms of both robustness and accuracy. This significant performance gain further supports the efficacy and adaptability of our approach, even when handling multi-modal data with diverse classes. The successful classification results achieved by our method on the multi-modal dataset highlight its potential in various real-world scenarios, where the dimensionality of sparse proportional data from different modalities need to be reduced to be effectively utilized for classification tasks. The ability of our method to consistently outperform the rest of the algorithms in this multi-modal data scenario further strengthens its position as a versatile and reliable solution for dimensionality reduction of highly sparse proportional data.

4.2. Complexity Analysis

In terms of performance and computational complexity, the proposed method solves a major problem in the compared algorithms. Most of the linear DR methods in the literature formulate and solve a GEVP. In cases of extremely high-dimensional data though, this GEVP cannot be solved efficiently. Moreover, often, solving this problem involves manipulating sparse high-dimensional matrices that result in near-singular matrices. The proposed method takes a different approach to resolve this issue. The original data are first projected into the new low-dimensional space, and, therefore, all the calculations are performed in the low-dimensional space, which avoids all the problems caused by the sparsity and high dimensionality of the data. In fact, the main reason SOLPP and LFDA did not yield any results in our experiments is that the data were extremely sparse. Also note that the proposed algorithm has the capacity to be parallelized easily, since the evaluation of the fitness function in the GA is independent for different members. This can be achieved even in the case of one processing unit leveraging vector calculations in software packages. In fact, the most intensive calculation in our proposed method is the matrix multiplication of the candidate projections with the data. This can be performed highly efficiently by concatenating all the projections, multiplying this matrix by the data, and then disassembling the resulting projected data. Such matrix multiplications are often performed very efficiently, with minimal time complexity in the current software libraries.
Moreover, the proposed method is significantly lighter in terms of resource usage since it (unlike the rest of the methods) does not need to invert any high-dimensional matrices that are of the order O ( N 3 ) , and, due to the fact that it only uses some statistics of the data (compared to the rest of the algorithms which have to build a neighborhood matrix of size M × M ), it needs much fewer resources, which becomes very important in the case that there are a large number of samples or the algorithm is running on low-power, low-resource devices. In general, the time complexity of solving the GEVP using dense linear algebra techniques is typically on the order of O ( N 3 ) , specifically O ( N 3 + c N 2 ) for some software libraries, where c is a constant depending on the specific implementation and machine architecture. If the matrix is sparse, specialized algorithms based on iterative methods are often used to solve the GEVP more efficiently. However, the time complexity of these sparse eigenvalue solvers is still on the order of O ( k N 3 ) to O ( k N 3 ) , where k is the number of desired eigenpairs (eigenvectors and eigenvalues), which, considering the large dimensionality of the data, is a significantly time-consuming process. Finally, the choice of the distributions used in the algorithm is affected by several properties of the data. In case there is no information about the underlying distribution of the data, the covariance matrix structure and the number of parameters of each distribution (in relation to the number of training samples) are good guides for choosing the possible candidate distributions.

5. Conclusions

High-dimensional sparse data are encountered more and more due to the low cost and ease of acquiring data, and several problems, including but not limited to the curse of dimensionality and resource usage, are associated with this type of data. In the present work, we have introduced an algorithm of supervised linear dimensionality reduction that remedies some of the problems of the conventional DR techniques in this context. The method has been tested against some of the well-known dimensionality reduction techniques and its superiority was demonstrated. Considering that the curse of dimensionality is a major problem while manipulating sparse data, our proposed method solves this problem effectively. The experiments show that EXMMP is, unlike the rest of the methods, robust in all cases of sparse data and yields better accuracies. It also can handle multi-modal data due to the fact that it models the data using a mixture of distributions, while the rest of the algorithms fail to do so. Moreover, the algorithm is scalable since increasing the number of features and samples have minimum effect on the complexity. Note that this increase only affects the first step of the algorithm, which is the projection into the low-dimensional space in the form of a matrix multiplication, while in the rest of the methods, it is carried along in the algorithm to the last step. Our contributions are briefly summarized below:
  • Proposing a novel supervised DR method that addresses the sparsity of the data effectively and efficiently.
  • Introducing a unique approach to handle multi-modal data by modeling the projected data using a mixture of exponential family distributions.
  • Formulating a closed form of KL-Divergence between the mixtures as a measure of separability.
Considering the above advantages, EXMMP is an effective tool in solving problems related to sparse high-dimensional data. Finally, as the next step in the further development of the algorithm, one can consider using other measures of distance between mixtures of distributions which yields better data separation. Also, the method can benefit from using a better approximation for the closed form of the KL-Divergence.

Author Contributions

Conceptualization, W.M. and N.B.; methodology, W.M.; software, W.M.; validation, W.M. and N.B.; data curation, W.M.; writing—original draft preparation, W.M.; writing—review and editing, W.M. and N.B.; visualization, W.M.; supervision, N.B.; funding acquisition, N.B. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Natural Sciences and Engineering Research Council of Canada (NSERC) under Grants to N. Bouguila (NSERC RGPIN/6656-2017).

Data Availability Statement

The datasets used for our experiments are publically available in the following links:

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Donoho, D.L. High-dimensional data analysis: The curses and blessings of dimensionality. AMS Math Chall. Lect. 2000, 1, 32. [Google Scholar]
  2. Pedagadi, S.; Orwell, J.; Velastin, S.; Boghossian, B. Local fisher discriminant analysis for pedestrian re-identification. In Proceedings of the 2013 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Washington, DC, USA, 23–28 June 2013; pp. 3318–3325. [Google Scholar]
  3. Chandrashekar, G.; Sahin, F. A survey on feature selection methods. Comput. Electr. Eng. 2014, 40, 16–28. [Google Scholar] [CrossRef]
  4. Liu, K.; Fu, Y.; Wu, L.; Li, X.; Aggarwal, C.; Xiong, H. Automated Feature Selection: A Reinforcement Learning Perspective. IEEE Trans. Knowl. Data Eng. 2023, 35, 2272–2284. [Google Scholar] [CrossRef]
  5. Bruni, V.; Cardinali, M.L.; Vitulano, D. A Short Review on Minimum Description Length: An Application to Dimension Reduction in PCA. Entropy 2022, 24, 269. [Google Scholar] [CrossRef] [PubMed]
  6. Abdulhammed, R.; Musafer, H.; Alessa, A.; Faezipour, M.; Abuzneid, A. Features Dimensionality Reduction Approaches for Machine Learning Based Network Intrusion Detection. Electronics 2019, 8, 322. [Google Scholar] [CrossRef] [Green Version]
  7. Chao, G.; Luo, Y.; Ding, W. Recent Advances in Supervised Dimension Reduction: A Survey. Mach. Learn. Knowl. Extr. 2019, 1, 341–358. [Google Scholar] [CrossRef] [Green Version]
  8. Cunningham, J.P.; Ghahramani, Z. Linear dimensionality reduction: Survey, insights, and generalizations. J. Mach. Learn. Res. 2015, 16, 2859–2900. [Google Scholar]
  9. Zhuo, L.; Cheng, B.; Zhang, J. A comparative study of dimensionality reduction methods for large-scale image retrieval. Neurocomputing 2014, 141, 202–210. [Google Scholar] [CrossRef]
  10. Lu, H.; Plataniotis, K.N.; Venetsanopoulos, A.N. A survey of multilinear subspace learning for tensor data. Pattern Recognit. 2011, 44, 1540–1551. [Google Scholar] [CrossRef]
  11. Jiang, X. Linear subspace learning-based dimensionality reduction. IEEE Signal Process. Mag. 2011, 28, 16–26. [Google Scholar] [CrossRef]
  12. Fisher, R.A. The use of multiple measurements in taxonomic problems. Ann. Eugen. 1936, 7, 179–188. [Google Scholar] [CrossRef]
  13. Jolliffe, I.T. Principal Component Analysis. Encycl. Stat. Behav. Sci. 2002, 30, 487. [Google Scholar]
  14. Sugiyama, M. Dimensionality reduction of multimodal labeled data by local fisher discriminant analysis. J. Mach. Learn. Res. 2007, 8, 1027–1061. [Google Scholar]
  15. Sugiyama, M.; Idé, T.; Nakajima, S.; Sese, J. Semi-supervised local Fisher discriminant analysis for dimensionality reduction. J. Mach. Learn. 2010, 78, 35–61. [Google Scholar] [CrossRef] [Green Version]
  16. Bartholomew, D.J. The foundations of factor analysis. Biometrika 1984, 71, 221–232. [Google Scholar] [CrossRef]
  17. He, X.; Niyogi, P. Locality preserving projections. Neural Inf. Process. Syst. 2004, 16, 153. [Google Scholar]
  18. Wang, H.; Lu, X.; Hu, Z.; Zheng, W. Fisher Discriminant Analysis With L1-Norm. IEEE Trans. Cybern. 2014, 44, 828–842. [Google Scholar] [CrossRef]
  19. Schölkopf, B.; Smola, A.; Müller, K.R. Nonlinear component analysis as a kernel eigenvalue problem. Neural Comput. 1998, 10, 1299–1319. [Google Scholar] [CrossRef] [Green Version]
  20. Weinberger, K.; Saul, L. An introduction to nonlinear dimensionality reduction by maximum variance unfolding. In Proceedings of the 2006 Twenty First National Conference on Artificial Intelligence, Boston, MA, USA, 16–20 July 2006; pp. 1683–1686. [Google Scholar]
  21. McInnes, L.; Healy, J.; Melville, J. UMAP: Uniform Manifold Approximation and Projection for Dimension Reduction. arXiv 2020, arXiv:1802.03426. [Google Scholar]
  22. van der Maaten, L.; Hinton, G. Visualizing Data using t-SNE. J. Mach. Learn. Res. 2008, 9, 2579–2605. [Google Scholar]
  23. Roweis, S.T.; Saul, L.K. Nonlinear dimensionality reduction by locally linear embedding. Science 2000, 290, 2323–2326. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  24. Hu, H.; Feng, D.; Yang, F. A Promising Nonlinear Dimensionality Reduction Method: Kernel-Based within Class Collaborative Preserving Discriminant Projection. IEEE Signal Process. Lett. 2020, 27, 2034–2038. [Google Scholar] [CrossRef]
  25. Bouguila, N.; Ziou, D. A dirichlet process mixture of generalized dirichlet distributions for proportional data modeling. IEEE Trans. Neural Netw. 2010, 21, 107–122. [Google Scholar] [CrossRef] [PubMed]
  26. Fan, W.; Bouguila, N. Learning finite Beta-Liouville mixture models via variational Bayes for proportional data clustering. In Proceedings of the 2013 IJCAI International Joint Conference on Artificial Intelligence, Beijing, China, 3–9 August 2013; pp. 1323–1329. [Google Scholar]
  27. Epaillard, E.; Bouguila, N. Proportional data modeling with hidden Markov models based on generalized Dirichlet and Beta-Liouville mixtures applied to anomaly detection in public areas. Pattern Recognit. 2016, 55, 125–136. [Google Scholar] [CrossRef]
  28. Masoudimansour, W.; Bouguila, N. Dimensionality reduction of proportional data through data separation using dirichlet distribution. Image Anal. Recognit. 2015, 9164, 141–149. [Google Scholar]
  29. Blei, D.M.; Ng, A.Y.; Jordan, M.I. Latent Dirichlet allocation. Mach. Learn. Res. 2012, 3, 993–1022. [Google Scholar]
  30. Wang, H.Y.; Yang, Q.; Qin, H.; Zha, H. Dirichlet component analysis: Feature extraction for compositional data. In Proceedings of the 2008 25th International Conference on Machine Learning, Helsinki, Finland, 5–9 July 2008; pp. 1128–1135. [Google Scholar]
  31. Epaillard, E.; Bouguila, N. Hidden Markov models based on generalized dirichlet mixtures for proportional data modeling. Lect. Notes Comput. Sci. 2014, 8774, 71–82. [Google Scholar]
  32. Shen, Z.H.; Pan, Y.H.; Wang, S.T. A supervised locality preserving projection algorithm for dimensionality reduction. Pattern Recognit. Artif. Intell. 2008, 21, 233–239. [Google Scholar]
  33. Wong, W.K.; Zhao, H.T. Supervised optimal locality preserving projection. Pattern Recognit. 2012, 45, 186–197. [Google Scholar] [CrossRef]
  34. Cai, D.; He, X.; Zhou, K.; Han, J.; Bao, H. Locality sensitive discriminant analysis. In Proceedings of the 20th International Joint Conference on Artifical Intelligence, Hyderabad, India, 6–12 January 2007; pp. 708–713. [Google Scholar]
  35. Kullback, S. Information Theory and Statistics; Dover Publications: Mineola, NY, USA, 1997. [Google Scholar]
  36. Hershey, J.R.; Olsen, P.A. Approximating the Kullback-Leibler divergence between Gaussian mixture models. Acoust. Speech Signal Process. 2007, 4, 317–320. [Google Scholar]
  37. Kuhn, H.W.; Tucker, A.W. Nonlinear programming. In Second Berkeley Symposium on Mathematical Statistics and Probability; Springer: Basel, Switzerland, 1951; pp. 481–492. [Google Scholar]
  38. Aitchison, J. The Statistical Analysis of Compositional Data; Monographs on Statistics and Applied Probability; Chapman and Hall: Boca Raton, FL, USA, 1986. [Google Scholar]
  39. Chaladze, G.; Kalatozishvili, L. Linnaeus 5 Dataset for Machine Learning. 2017. Available online: http://chaladze.com/l5/ (accessed on 24 June 2023).
Figure 1. Sample images from different categories (classes) in Caltech-256 dataset.
Figure 1. Sample images from different categories (classes) in Caltech-256 dataset.
Electronics 12 03355 g001
Table 1. Classification accuracy (%) of 20-newsgroups dataset for different target dimensions K, and 1000 original features. Boldface shows the best result.
Table 1. Classification accuracy (%) of 20-newsgroups dataset for different target dimensions K, and 1000 original features. Boldface shows the best result.
K3456
SLPP78.33 ± 0.1178.27 ± 0.1277.52 ± 0.1077.85 ± 0.11
LSDA78.22 ± 0.1077.47 ± 0.1078.81 ± 0.1177.96 ± 0.12
LDA-L182.62 ± 0.0382.74 ± 0.0480.81 ± 0.0283.48 ± 0.01
EXPMMP94.15 ± 0.0195.00 ± 0.0194.89 ± 0.0194.94 ± 0.01
Table 2. Classification accuracy (%) of Caltech-256 dataset for unimodal data, K = 3 . Boldface shows best result.
Table 2. Classification accuracy (%) of Caltech-256 dataset for unimodal data, K = 3 . Boldface shows best result.
ClassesSLPPLSDALDA-L1EXPMMP
Greyhound vs. Galaxy92.60 ± 0.0492.75 ± 0.0490.90 ± 0.0394.90 ± 0.02
Lightning vs. Hourglass81.02 ± 0.0381.42 ± 0.0372.38 ± 0.0384.15 ± 0.02
Eyeglasses vs. Microscope81.00 ± 0.0381.00 ± 0.0377.50 ± 0.0383.00 ± 0.02
Hrmonica vs. Laptop71.38 ± 0.0371.51 ± 0.0367.71 ± 0.0276.03 ± 0.02
Necktie vs. Conch71.77 ± 0.0571.75 ± 0.0673.50 ± 0.0578.07 ± 0.02
Gas pump vs. Yarmulke76.55 ± 0.0676.13 ± 0.0665.95 ± 0.0483.30 ± 0.03
Kayak vs. Chandelier77.52 ± 0.0377.35 ± 0.0364.55 ± 0.0478.95 ± 0.02
Unicorn vs. Lathe73.69 ± 0.0373.91 ± 0.0463.78 ± 0.0375.26 ± 0.01
Motorbikes vs. Breadmaker91.70 ± 0.0291.37 ± 0.0289.04 ± 0.0197.12 ± 0.01
Table 3. Classification accuracy (%) of Linnaeus-5 dataset for unimodal data, K = 3 . Boldface shows best result.
Table 3. Classification accuracy (%) of Linnaeus-5 dataset for unimodal data, K = 3 . Boldface shows best result.
ClassesSLPPLSDALDA-L1EXPMMP
Berry vs. Bird76.18 ± 0.0376.25 ± 0.0368.06 ± 0.0377.62 ± 0.02
Berry vs. Dog77.68 ± 0.0577.37 ± 0.0468.09 ± 0.0478.50 ± 0.01
Berry vs. Flower68.90 ± 0.0369.34 ± 0.0462.34 ± 0.0369.93 ± 0.02
Berry vs. Other79.96 ± 0.0479.46 ± 0.0470.96 ± 0.0382.78 ± 0.02
Bird vs. Dog63.84 ± 0.0364.09 ± 0.0358.25 ± 0.0367.56 ± 0.01
Bird vs. Flower71.37 ± 0.0471.31 ± 0.0462.34 ± 0.0369.90 ± 0.01
Bird vs. Other67.06 ± 0.0466.90 ± 0.0459.43 ± 0.0469.68 ± 0.03
Dog vs. Flower74.15 ± 0.0374.28 ± 0.0261.46 ± 0.0375.12 ± 0.01
Dog vs. Other74.59 ± 0.0375.00 ± 0.0368.75 ± 0.0377.75 ± 0.02
Flower vs. Other76.59 ± 0.0376.65 ± 0.0268.56 ± 0.0279.03 ± 0.01
Table 4. Classification accuracy (%) of 20-newsgroups dataset for unimodal data, K = 3 . Boldface shows best result.
Table 4. Classification accuracy (%) of 20-newsgroups dataset for unimodal data, K = 3 . Boldface shows best result.
ClassesSLPPLSDALDA-L1EXPMMP
Hockey vs. Crypt90.84 ± 0.0291.58 ± 0.0280.61 ± 0.0493.85 ± 0.01
Electronics vs. Religion75.72 ± 0.0275.66 ± 0.0273.24 ± 0.0390.19 ± 0.01
Politics vs. Hockey87.52 ± 0.0187.63 ± 0.0177.64 ± 0.0492.88 ± 0.01
Christian vs. Religion66.19 ± 0.0165.95 ± 0.0167.79 ± 0.0276.11 ± 0.01
Windows vs. Electronics74.46 ± 0.0880.36 ± 0.0265.20 ± 0.0483.11 ± 0.01
Graphics vs. Guns86.32 ± 0.0186.26 ± 0.0176.74 ± 0.0191.37 ± 0.01
Autos vs. Politics81.43 ± 0.0181.54 ± 0.0169.84 ± 0.0487.56 ± 0.01
Atheism vs. Baseball85.63 ± 0.0285.13 ± 0.0171.88 ± 0.0187.81 ± 0.01
Table 5. Classification accuracy (%) of ImageNet dataset for unimodal data, K = 3 . Boldface shows best result.
Table 5. Classification accuracy (%) of ImageNet dataset for unimodal data, K = 3 . Boldface shows best result.
ClassesSLPPLSDALDA-L1EXPMMP
statue vs. azure65.03 ± 0.0364.98 ± 0.0260.67 ± 0.0370.13 ± 0.02
outbuilding vs. high altar78.00 ± 0.0277.54 ± 0.0267.46 ± 0.0381.68 ± 0.02
falcon vs. leash62.63 ± 0.0263.71 ± 0.0264.04 ± 0.0272.52 ± 0.02
blue peafowl vs. horseman71.23 ± 0.0271.48 ± 0.0267.41 ± 0.0177.18 ± 0.01
purl vs. odometer89.77 ± 0.0289.72 ± 0.0283.52 ± 0.0393.50 ± 0.02
chickadee vs. pipe71.84 ± 0.0272.14 ± 0.0272.14 ± 0.0277.27 ± 0.01
woodpecker vs. kitten73.37 ± 0.0273.12 ± 0.0368.02 ± 0.0380.68 ± 0.02
watermelon vs. crocodile71.07 ± 0.0171.32 ± 0.0169.54 ± 0.0280.30 ± 0.01
guinea pig vs. sea eagle72.69 ± 0.0373.14 ± 0.0266.68 ± 0.0274.19 ± 0.02
caiman vs. parrotfish67.21 ± 0.0266.70 ± 0.0370.35 ± 0.0279.13 ± 0.02
fig vs. wild carrot85.41 ± 0.0285.14 ± 0.0281.62 ± 0.0391.43 ± 0.01
Table 6. Classification accuracy (%) of Food-101 dataset for unimodal data, K = 3 . Boldface shows best result.
Table 6. Classification accuracy (%) of Food-101 dataset for unimodal data, K = 3 . Boldface shows best result.
ClassesSLPPLSDALDA-L1EXPMMP
carrot cake vs. poutine76.75 ± 0.0276.65 ± 0.0271.40 ± 0.0384.15 ± 0.02
miso soup vs. pad thai89.95 ± 0.0290.25 ± 0.0285.10 ± 0.0294.45 ± 0.02
panna cotta vs. ramen75.30 ± 0.0274.05 ± 0.0269.65 ± 0.0479.90 ± 0.01
cheesecake vs. tacos76.40 ± 0.0276.30 ± 0.0274.55 ± 0.0282.50 ± 0.02
ice cream vs. tuna tartare65.25 ± 0.0364.65 ± 0.0362.55 ± 0.0374.20 ± 0.03
fried rice vs. waffles81.85 ± 0.0281.30 ± 0.0278.10 ± 0.0287.85 ± 0.02
caesar salad vs. oysters74.85 ± 0.0374.95 ± 0.0370.50 ± 0.0183.40 ± 0.01
hot and sour soup vs. sushi83.35 ± 0.0283.15 ± 0.0278.95 ± 0.0291.20 ± 0.01
Table 7. Classification accuracy (%) of Food-101 dataset for multi-modal data, K = 3 . Boldface shows best result.
Table 7. Classification accuracy (%) of Food-101 dataset for multi-modal data, K = 3 . Boldface shows best result.
ClassesSLPPLSDALDA-L1EXPMMP
french onion soup + garlic bread
vs. strawberry shortcake
77.16 ± 0.0276.60 ± 0.0267.80 ± 0.0279.83 ± 0.02
breakfast burrito + cheesecake
vs. miso soup
83.63 ± 0.0283.33 ± 0.0275.53 ± 0.0186.13 ± 0.01
french toast + greek salad
vs. seaweed salad
82.56 ± 0.0182.30 ± 0.0173.26 ± 0.0284.80 ± 0.01
breakfast burrito + hummus
vs. ramen
75.76 ± 0.0375.63 ± 0.0368.40 ± 0.0278.46 ± 0.02
baklava + fish and chips
vs. strawberry shortcake
72.46 ± 0.0271.90 ± 0.0267.36 ± 0.0274.23 ± 0.02
cheese plate + pancakes
vs. spring rolls
73.40 ± 0.0273.50 ± 0.0268.26 ± 0.0277.90 ± 0.02
chicken curry + crab cakes
vs. foie gras
70.26 ± 0.0270.80 ± 0.0269.26 ± 0.0176.02 ± 0.01
beef tartare + french toast
vs. spaghetti bolognese
82.60 ± 0.0282.73 ± 0.0173.93 ± 0.0185.20 ± 0.01
beef tartare + cup cakes
vs. lobster bisque
84.43 ± 0.0284.43 ± 0.0278.13 ± 0.0187.26 ± 0.01
donuts + gyoza
vs. pho
86.43 ± 0.0185.96 ± 0.0176.56 ± 0.0188.10 ± 0.01
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Masoudimansour, W.; Bouguila, N. Supervised Dimensionality Reduction of Proportional Data Using Exponential Family Distributions. Electronics 2023, 12, 3355. https://doi.org/10.3390/electronics12153355

AMA Style

Masoudimansour W, Bouguila N. Supervised Dimensionality Reduction of Proportional Data Using Exponential Family Distributions. Electronics. 2023; 12(15):3355. https://doi.org/10.3390/electronics12153355

Chicago/Turabian Style

Masoudimansour, Walid, and Nizar Bouguila. 2023. "Supervised Dimensionality Reduction of Proportional Data Using Exponential Family Distributions" Electronics 12, no. 15: 3355. https://doi.org/10.3390/electronics12153355

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop