 Previous Article in Journal
Modernising Receiver Operating Characteristic (ROC) Curves

Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

# Well-Separated Pair Decompositions for High-Dimensional Datasets

Department of Mathematics, University of Osijek, 31000 Osijek, Croatia
Algorithms 2023, 16(5), 254; https://doi.org/10.3390/a16050254
Received: 4 April 2023 / Revised: 9 May 2023 / Accepted: 11 May 2023 / Published: 15 May 2023

## Abstract

:
Well-separated pair decomposition (WSPD) is a well known geometric decomposition used for encoding distances, introduced in a seminal paper by Paul B. Callahan and S. Rao Kosaraju in 1995. WSPD compresses $O ( n 2 )$ pairwise distances of n given points from $R d$ in $O ( n )$ space for a fixed dimension d. However, the main problem with this remarkable decomposition is the “hidden” dependence on the dimension d, which in practice does not allow for the computation of a WSPD for any dimension $d > 2$ or $d > 3$ at best. In this work, I will show how to compute a WSPD for points in $R d$ and for any dimension d. Instead of computing a WSPD directly in $R d$, I propose to learn nonlinear mapping and transform the data to a lower-dimensional space $R d ′$, $d ′ = 2$ or $d ′ = 3$, since only in such low-dimensional spaces can a WSPD be efficiently computed. Furthermore, I estimate the quality of the computed WSPD in the original $R d$ space. My experiments show that for different synthetic and real-world datasets my approach allows that a WSPD of size $O ( n )$ can still be computed for points in $R d$ for dimensions d much larger than two or three in practice.

## 1. Introduction

Given a set of points $S ⊂ R d$, $| S | = n$, a well-separated pair decomposition (WSPD) can be seen as a compressed representation for approximating $n 2$ pairwise distances of n points from S into $O ( n )$ space, where the dimension d is considered a constant. The formal definition of a WSPD will be given in Section 2. A WSPD can also be seen as a clustering approach, e.g., a WSPD is a partition of the $n 2$ edges of the complete Euclidean graph into $O ( n )$ subsets. This decomposition was first introduced in a seminal paper  by Paul B. Callahan and S. Rao Kosaraju in 1995. It has been shown in  that the size of a WSPD, when computed by their algorithm, is exponential in d. Hence, it has never been used in practice for dimensions larger than three.
However, a WSPD has been shown to be useful in many different applications. It is known that a WSPD can be used to efficiently solve a number of proximity problems , such as the closest pair problem, the all-nearest neighbors problem, etc. It is also known that a WSPD directly induces a t-spanner of a point set or provides a $( 1 + ε )$ —approximation of the Euclidean minimum spanning tree. The authors in  used a WSPD to compute approximate energy-efficient paths in radio networks. Their algorithm reduced the complexity of computing such paths to $O ( 1 )$ by moving most of the computation to the preprocessing stage by precomputing a template path for each pair of sets compressing the pairwise distances.
Since a WSPD proved to be an essential decomposition for different important problems, in this paper, I investigate to what extent a WSPD can be helpful for situations where the input $S ⊂ R d$ and the dimension d is much larger than two or three.
On the technical side, a WSPD of a set of points $S ⊂ R d$ is represented by a sequence of pairs of sets $( A i , B i )$, $i = 1 , … , k$, called dumbbells, such that (i) for every two distinct points $a , b ∈ S$ there exists a unique dumbbell $( A i , B i )$ such that $a ∈ A i , b ∈ B i$; (ii) the distances between points in $A i$ and $B i$ are approximately equal; (iii) the distances between points in $A i$ or points in $B i$ are much smaller than the distances between points in $A i$ and $B i$.
As stated before, the size of a WSPD, i.e., the number of dumbbells, is known to grow exponentially with the dimension d. Hence, instead of computing a WSPD directly on $S ⊂ R d$, I first propose to transform the points in S with a nonlinear function $f θ : R d → R d ′$, $d ′ = 2$ or $d ′ = 3$, where $θ$ denotes the set of learnable function parameters. The parameters $θ$ are determined such that the function $f θ$ preserves the properties of S that are important for a WSPD, e.g., preserves pairwise distances for points in S. If the function $f θ$ manages to preserve most of the important information for a WSPD, there is hope that the WSPD computed on the mapped points $f θ ( S ) ⊂ R d ′$, $d ′ = 2$ or $d ′ = 3$, where the size of the WSPD is $O ( n )$, will output “dumbbells” that are meaningful even for the original input S. If this is the case, the reconstructed dumbbells should continue to be “dumbbell-shaped” in the original space. In practice, some of the reconstructed dumbbells can become “bad” because they do not approximate the distances in any practical sense. However, I will show that the number of “bad” dumbbells is negligible in practice. Moreover, such “bad” dumbbells can be easily refined without significantly increasing the total number of dumbbells. One tool I employ might be of independent interest: I implemented a WSPD following the nontrivial algorithm of the partial fair-split tree that guarantees the construction time of $O ( n log n )$ and a WSPD of $O ( n )$ size. To my knowledge, my implementation is the first open-source publicly available implementation of a WSPD that carefully follows the original algorithm in . The implementation of a WSPD in the ParGeo C++ library (see ) uses a simple fair-split kd-tree and as such does not ensure theoretical bounds on the size of a WSPD.
Recently, there have been attempts to improve classical clustering for high-dimensional datasets. The authors in  have proposed a deep embedded clustering method that simultaneously learns feature representations and cluster assignments using a nonlinear mapper $f θ$. The work of [6,7] proposed an interesting Maximal Coding Rate Reduction principle for determining the parameters $θ$ of the function $f θ$. The work of  further developed that idea in the context of manifold clustering.
Although this work was motivated by the research above, I note that their use of a function $f θ$ was aimed at improving clustering in the sense that a nonlinear mapper $f θ$ was used as an additional mechanism to “learn” a better set of features that would enable a better clustering. In my approach, $f θ$, which is represented by a neural network, is primarily used as a mapper to a very low-dimensional representation of the original dataset, since only there can a WSPD be computed efficiently.
In the rest of this paper, I will formally define a WSPD and state two important theorems. Furthermore, I will introduce two different functions for $f θ$ and present steps for computing a WSPD for high-dimensional data sets. Finally, I provide empirical evidence for the claim that a WSPD of size $O ( n )$ can be computed efficiently for dimensions d much larger than three.

## 2. Preliminaries

Let S be a set of n points in $R d$. For any $A ⊆ S$, let $R ( A )$ denote the minimum enclosing axis-aligned box of A. Let $C A$ be the minimum enclosing ball of $R ( A )$, and let $r ( A )$ denote the radius of $C A$. Let $C A r$ be the ball with the same center as $C A$ but with radius r. Furthermore, for two sets A, $B ⊆ S$, let $r = max ( r ( A ) , r ( B ) )$, and let $d ( A , B )$ denote the minimum distance between $C A r$ and $C B r$. For example, if the $C A$ intersects $C B$, then $d ( A , B ) = 0$ (Figure 1).
Definition 1.
A pair of sets A and B are said to be well separated (a dumbbell) if $d ( A , B ) > s · r$, for any given separation constant $s > 0$ and $r = max { r ( A ) , r ( B ) }$.
Definition 2
(WSPD). A well-separated pair decomposition of $S ⊂ R d$, for a given $s > 0$, is a sequence $( A 1 , B 1 ) , … , ( A k , B k )$, where $A i , B i ⊆ S$, such that the following applies:
1.
$A i , B i$ are well separated with respect to separation constant s, for all $i = 1 , … , k$;
2.
For all $p ≠ q ∈ S$, there exists a unique pair $( A i , B i )$ such that $p ∈ A i , q ∈ B i$ or $q ∈ A i , p ∈ B i$.
Note that a WSPD always exists since one could use all singleton pairs $( { p } { q } )$, for all pairs $p , q ∈ S$. However, this would yield a sequence of dumbbells of size $k = Θ ( n 2 )$. The question is whether one could do better than that. The answer to that question was given by the following theorem.
Theorem 1
(). Given a set S of n points in $R d$ and a separation constant $s > 0$, a WSPD of S with $O ( s d d d / 2 n )$ amount of dumbbells can be computed in $O ( d n log n + s d d d / 2 n )$ time.
While Theorem 1 states that it is possible to compute only $O ( n )$ dumbbells, for fixed dimension d, it is still not clear how to efficiently determine the appropriate dumbbell for a given query pair $( p , q )$. In , it has been shown that retrieving the corresponding dumbbell can be performed in $O ( 1 )$ time for a fixed dimension d.
Theorem 2
(). Given a well-separated pair decomposition of a point set S with separation constant $s > 2$ and fixed dimension d, I can construct a data structure in space $O ( n · s 2 )$ and construction time $O ( n · s 2 )$ such that for any pair of points $( p , q )$ in S I can determine the unique pair of clusters $( A , B )$ that is part of the well-separated pair decomposition with $p ∈ A , q ∈ B$ in constant time.

## 3. Deep Embedded WSPD

Instead of computing a WSPD of $S ⊂ R d$, as suggested by Theorem 1, I propose to first transform the data with a nonlinear mapping $f θ : R d → R d ′$, where $d ′$ is chosen to be either 2 or 3, and $θ$ is a set of learnable parameters. In this section, I introduce two approaches for training the neural net for embedding the point sets.

#### 3.1. Metric Multidimensional Scaling

The most natural choice for a function $f θ$ is to use metric multidimensional scaling (mMDS), which tries to preserve the pairwise distances between points, i.e., it solves the optimization problem
$min θ ∑ i , j w i j ( | | x i − x j | | − | | f θ ( x i ) − f θ ( x j ) | | ) 2 ,$
where $| | · | |$ denotes the Euclidean norm and $w i j ≥ 0$ are some given weights. For the implementation of a function $f θ$, I will use deep neural networks. The neural network approach for computing the metric MDS mapper has already been used in . Inspired by the work of  [5,10,11], I chose the following architecture with $L =$ five fully connected layers of the form below:
$d − 500 − 500 − 2000 − d ′$
and $d ′ ∈ { 2 , 3 }$. Moreover, the output $h k x$, $k > 1$ of the kth layer for each $x ∈ S$ is defined as follows:
$h k x = g ( W k − 1 h k − 1 x + b k − 1 )$
where $h 1 x$ is just another name for any $x ∈ S$, $g ( · )$ denotes an activation function and $θ = { W k , b k | k = 1 , … , L − 1 }$ are model parameters. For the activation function $g ( · )$, I use the hyperbolic tangent (tanh).

#### 3.2. Autoencoder

We will also implement the function $f θ$ as a stacked autoencoder, a type of neural network typically used to learn encodings of unlabeled data. It is known that such a learned data representation maintains semantically meaningful information ([12,13]). Moreover, it has been shown in , one of the seminal papers in deep learning, that an autoencoder can be effectively used on real-world datasets as a nonlinear generalization of the widely used principal component analysis (PCA).
The fundamental concept is to use an encoder to reduce high-dimensional data to a low-dimensional space. However, this results in a loss of information in the data. The decoder then works to map the data back to its original space. The better the mapping, the less information is lost in the process. Thus, the basic idea of an autoencoder is to have an output layer with the same dimensionality as the inputs. In contrast, the number of units in the middle layer is typically much smaller compared to the inputs or outputs. Therefore, it is assumed that the middle layer units contain a reduced data representation. Since the output is supposed to approximate the input, it is hoped that the reduced representation preserves “interesting” properties of the input (e.g., pairwise distances). It is common that autoencoders have a symmetric architecture, i.e., for an odd number L, the number of units in the kth layer of an L-layer autoencoder architecture is equal to the number of units in the $( L − k + 1 )$th layer. Moreover, the first part of the network, up to the middle layer, is called the encoder, while the part from the middle layer to the outputs is called the decoder.
We use a very similar architecture as above, namely $L =$ nine layers of the form below:
$d − 500 − 500 − 2000 − d ′ − 2000 − 500 − 500 − d$
for $d ′ ∈ { 2 , 3 }$. The output $h k x$, $k > 1$ of the kth layer for any $x ∈ S$ is computed as above using the tanh activation function. Training is conducted by minimizing the following least squares loss:
$min θ ∑ x ∈ S | | x − h L x | | 2 .$
Once the autoencoder is trained on a given dataset S, the encoder part of the network $d − 500 − 500 − 2000 − d ′$ is used as a nonlinear mapper $f θ : R d → R d ′$, and the decoder part is discarded.

#### 3.3. Computing WSPD in High Dimensions

Given a function $f θ$, the algorithm for computing a WSPD of a high-dimensional dataset $S ⊂ R d$ is given by the following steps:
• Compute a lower-dimensional representation $S ′ = f θ ( S )$.
• Compute a WSPD$( S ′ , s )$, for any given $s > 0$, as proposed by Theorem 1. Let $( A i ′ , B i ′ )$, $i = 1 , … , k$, denote the dumbbells.
• Reconstruct $( A i , B i )$, $i = 1 , … , k$, in the original $R d$ space. Note that not all of them are well separated, i.e., they are not dumbbells anymore.
• Refine all not well-separated pairs $( A i , B i )$ until they become dumbbells.
We will refer to the above steps as the NN-WSPD algorithm. Note that the reconstruction step in NN-WSPD can be performed efficiently. What requires further explanation is the refine step, which guarantees that the set of computed dumbbells in $R d$ is in fact again a well separated pair decomposition of S. Suppose $( A i , B i )$ for some i is not well separated and let $r ( B i ) > r ( A i )$. We propose the following steps in Algorithm 1 to refine $( A i , B i )$.
 Algorithm 1 Refine($A i$, $B i$) if $( A i , B i )$ is not a dumbbell then    Split $B i$ into $B i ′$ and $B i ″$    ▹Assuming $r ( B i ) > r ( A i )$, otherwise split $A i$    Remove $( A i , B i )$ from WSPD.    Add $( A i , B i ′ )$ and $( A i , B i ″ )$ to WSPD    Refine($A i$, $B i ′$) and Refine($A i , B i ″$)           ▹(Two recursive calls)end if
The complexity of the refine step depends on the recursion depth. We will experimentally demonstrate that the depth is relatively small for all the datasets that I tried in practice, introducing just a moderate number of new dumbbells.

#### 3.4. Fast Dumbbell Retrieval in High Dimensions

As Theorem 2 stated, for any pair of points $a , b ∈ S$ I can determine the unique dumbbell $( A , B )$, $a ∈ A , b ∈ B$, in constant time but only if dimension d is considered constant. The hidden constant in the running time is again exponential in d (due to the packing argument used in , Lemma 10). Hence, the only way for a pair of points $( a , b ) ∈ S$ to retrieve the corresponding dumbbell $( A , B )$ efficiently is to build and query the data structure proposed in  in lower-dimensional representation $R d ′ , d ′ = 2 , 3$. Namely, let $( A i ′ , B i ′ ) , i = 1 , … , k ′$, denote the dumbbells in $R d ′$ and $( A i , B i ) , i = 1 , … , k$, the reconstructed and refined dumbbells in $R d$ computed by NN-WSPD. Note that $k ≥ k ′$ in general, since the number of reconstructed and refined dumbbells might be larger. However, let $( f θ ( A i ) , f θ ( B i ) )$, $i = 1 , … , k$, denote the corresponding dumbbells in $R d ′$ of the WSPD computed by NN-WSPD in $R d$. Note that $( f θ ( A i ) , f θ ( B i ) )$, $i = 1 , … , k$ is also a valid WSPD of $S ′ = f θ ( S )$, and let query$( · , · )$ denote the query call to the data structure proposed in  built on that WSPD. The query algorithm for any two points $( a , b ) ∈ S ⊂ R d$ is defined in Algorithm 2.
 Algorithm 2 RetrieveDumbbell(a, b) Require:  $f θ$, $( f θ ( A i ) , f θ ( B i ) )$ for $i = 1 , … , k$, query$( · , · )$  Let $a ′ = f θ ( a )$, $b ′ = f θ ( b )$  $( A ′ , B ′ )$ = query$( a ′ , b ′ )$       ▹ Retrieve dumbbells by query$( · , · )$ from   Return $( A , B )$ such that $A ′ = f θ ( A ) , B ′ = f θ ( B )$
The run-time complexity of Algorithm 2 again depends on the number of additional dumbbells that the refine step will add.

## 4. Experiments

Our implementations of the neural networks are performed in PyTorch , and a WSPD is implemented in C++ following the algorithm proposed in . All codes are available for testing at https://github.com/dmatijev/high_dim_wspd.git. The experiments were conducted on Ryzen 9 3900X, with 12 cores and 64GB of DDR4 RAM with an Ubuntu 20.04 operating system.
For the training of the neural networks, I use an initial learning rate of 0.001 and a batch size of 512. All networks are trained for a number of epochs set to 500. It is essential to state that I did not put effort into fine-tuning these hyperparameters. Instead, all hyperparameters are set to achieve a reasonably good WSPD reconstruction and, to maintain fairness, are held constant across all datasets. We used Adam  for first-order gradient-based optimization and the ReduceLROnPlateau callback for reducing the learning rate (dividing it by 10) when a metric, given by (1) for mMDS NN, and (3) in the case of autoencoder NN, has stopped improving.

#### 4.1. Datasets

We evaluated the computation of a WSPD on artificially generated and real high-dimensional datasets. Artificially generated datasets are drawn from the following:
• A uniform distribution over the $[ 0 , 1 ] d$ hypercube;
• A normal (Gaussian) distribution, with mean 0 and standard deviation 1;
• A Laplace or double exponential distribution, with the position of the distribution peak at 0 and the exponential decay at 1.
For real-world datasets, I used two public scRNA-seq datasets downloaded from . scRNA-seq data are used to assess which genes are turned on in a cell and in what amount. Therefore, scRNA-data are typically used in computational biology to determine transcriptional similarities and differences within a population of cells, allowing for a better understanding of the biology of a cell. We apply the standard preprocessing to scRNA-seq data (see [18,19]): (a) compute the natural log-transformation of gene counts after adding a pseudo count of 1 and (b) select the top 2000 most variable genes followed by a dimensionality reduction to 100 principle components. We used two datasets of sizes $n = 4185$ and $n = 68 , 575$.
One possible motivation for a WSPD on scRNA-seq data can be found in the work of . Namely, the authors in  solve the marker gene selection problem by solving a linear program (LP). For a large amount of data, the LP cannot be efficiently solved in practice. Hence, the practical approach to that issue could be to solve an LP on a subset of constraints, then define a separation oracle that iteratively introduces new constraints to the LP, if the current solution is not feasible. Following the work of , the oracle could add the currently most violated constraints, i.e., pairs of points whose distance is less than a predefined constant value. Given the WSPD, one could speed up the oracle by using dumbbells as a substitute for the pairwise distances.

#### 4.2. Measuring the Quality of a WSPD

Let $( A i , B i )$, $i = 1 , … , k$, denote the WSPD for some set of points $S ⊂ R d$. For $a , a ′ ∈ A i$ and $b , b ′ ∈ B i$ for some dumbbell i, I make the following observations:
• Points within the sets $A i$ and $B i$ can be made “arbitrarily close” as compared to points in the opposite sets by choosing the appropriate separation $s > 0$, i.e.,
$d ( a , a ′ ) ≤ 2 r < 2 s d ( A i , B i ) ≤ 2 s d ( a , b ) .$
• Distances between points in the opposite sets can be made “almost equal”, by choosing the appropriate $s > 0$, i.e.,
$d ( a ′ , b ′ ) ≤ d ( a , a ′ ) + d ( a , b ) + d ( b , b ′ ) < ( 1 + 4 s ) d ( a , b ) .$
Since a WSPD is primarily concerned with compressing the quadratic space of pairwise distances into a linear space (for a fixed dimension d), I will use Equation (5) repeatedly in my plots in order to measure how much distances are indeed preserved within a dumbbell.

#### 4.3. Results

In Figure 2 (left), synthetic datasets were used to demonstrate the dependence between the size of a WSPD and the dimension of the input data in practice. Recall that the number of dumbbells is bounded above by $O ( s d d d / 2 n )$ (Theorem 1). In my experiments, I found that the dependence on dimension d is indeed severe, making the WSPD algorithm proposed in  unusable in practice for dimensions $d > 2$ or $d > 3$. However, it is unclear whether the number of dumbbells is just an artifact of the construction of the WSPD algorithm or whether that number of dumbbells is indeed necessary to satisfy the properties of a WSPD given by Definition 2. Thus, Figure 2 (right) demonstrates the total number of dumbbells when computed with the algorithm proposed in  and compares it to the number of dumbbells computed with the NN-WSPD approach.
• Observation: For many practical datasets, there exists a WSPD $( A i , B i )$, $i = 1 , … , k$, with $k = O ( n )$ for any dimension d, i.e., the hidden constant in $O ( · )$ notation is not exponential in the dimension d.
I performed numerous experiments with synthetic data and my two real datasets to support my claim.

#### 4.3.1. Synthetic Datasets

In Figure 3, the experiments were performed with the synthetic datasets. Only for $d = 2$ was the standard WSPD algorithm used, and for $d > 2$, NN-WSPD was used. On the boxplots in the left column, NN-WSPD was applied but without the refine step (Algorithm 1). From this one can see that many of the reconstructed dumbbells can violate the quality given by Equation (5), which the dumbbells are supposed to guarantee. However, most dumbbells are still well separated since a lot of valuable information for a WSPD was preserved by the nonlinear mapper $f θ$. The results of the refine step (Algorithm 1) can be seen in boxplots in the right column in Figure 3. Note that after the refine step all dumbbells are indeed dumbbells (i.e., well separated), with the overall number of dumbbells that rarely gets larger than the starting WSPD size multiplied by a factor of at most three, independent of the dimension d. We noticed that the higher the dimension d of the input set S, the fewer refined dumbbells are needed. This is not that surprising bearing in mind the fact that when a Euclidean distance is defined using many coordinates there is less difference in the distances between different pairs of samples (see the curse of dimensionality phenomenon ).

#### 4.3.2. scRNA-Seq Datasets

I had even fewer problems computing a WSPD for my two real datasets ($n = 4185$ and n = 68,579) that, after being preprocessed, were given in $R d$, $d = 100$. In Figure 4, I output boxplots for both sets before and after the refining step. Note that the number of newly added refined dumbbells is negligible, even compared to the original sets’ size. Moreover, notice that computing a WSPD directly on such a high-dimensional input ($d = 100$) is very inefficient and practically of no use due to the dependence on d. For example, I managed to compute a WSPD for dataset $n = 4185$ in 507 s with 8,401,551 dumbbells, which is slightly above 95% of the overall size of pairwise distances, i.e., 95% of dumbbells were just singletone pairs $( { a } , { b } )$. In contrast, my NN-WSPD approach always outputs a WSPD with a very moderate increase in size compared to a WSPD of size $O ( n )$ computed in the plane the points were projected to, as my experiments showed.

## 5. Conclusions

Well-separated pair decomposition is a well known decomposition in computational geometry. However, computational geometry as a field is concerned with algorithms that solve problems on datasets in $R d$, where the dimension d is considered a constant. In this work, I demonstrated that a WSPD of size $O ( n )$ could be computed even for high-dimensional sets, hence removing the requirement that dimension d is a constant. In my approach, I used an implementation of a nonlinear function $f θ$ that was based on artificial neural networks.
The past decade has seen remarkable advances in deep learning approaches based on artificial neural networks. We also witnessed a few successful applications of neural networks in discrete problems, e.g., I would like to point the reader to the surprising results presented in , which introduces a new neural network architecture (Ptr-Net) and shows that it can be used to learn approximate solutions to well-known geometric problems, such as planar convex hulls, Delaunay triangulations, and the planar traveling salesperson problem.
These advances inspire us to further explore the applications of deep learning in fields such as geometry.

## Funding

This research received no external funding.

## Data Availability Statement

Two public single-cell scRNA-seq datasets can be downloaded from . Codes are available at https://github.com/dmatijev/high_dim_wspd.git.

## Acknowledgments

I thank Nathan Chappell and Tomislav Prusina, who provided insight and expertise that greatly assisted this research.

## Conflicts of Interest

The author declares no conflict of interest.

## References

1. Callahan, P.B.; Kosaraju, S.R. A Decomposition of Multidimensional Point Sets with Applications to K-Nearest-Neighbors and n-Body Potential Fields. J. ACM 1995, 42, 67–90. [Google Scholar] [CrossRef]
2. Smid, M.H.M. The Well-Separated Pair Decomposition and Its Applications. In Handbook of Approximation Algorithms and Metaheuristics; Chapman and Hall/CRC: Boca Raton, FL, USA, 2007. [Google Scholar] [CrossRef]
3. Beier, R.; Funke, S.; Matijević, D.; Sanders, P. Energy-Efficient Paths in Radio Networks. Algorithmica 2011, 61, 298–319. [Google Scholar] [CrossRef]
4. Wang, Y.; Yu, S.; Dhulipala, L.; Gu, Y.; Shun, J. ParGeo: A Library for Parallel Computational Geometry. In Proceedings of the 27th ACM SIGPLAN Symposium on Principles and Practice of Parallel Programming, Seoul, Republic of Korea, 2–6 April 2022; Association for Computing Machinery: New York, NY, USA, 2022; pp. 450–452. [Google Scholar] [CrossRef]
5. Xie, J.; Girshick, R.; Farhadi, A. Unsupervised Deep Embedding for Clustering Analysis. In Machine Learning Research, Proceedings of the 33rd International Conference on Machine Learning, New York, NY, USA, 19–24 June 2016; PMLR: New York, NY, USA, 2016; Volume 48, pp. 478–487. [Google Scholar]
6. Yu, Y.; Chan, K.H.R.; You, C.; Song, C.; Ma, Y. Learning Diverse and Discriminative Representations via the Principle of Maximal Coding Rate Reduction. In Advances in Neural Information Processing Systems; Curran Associates, Inc.: New York, NY, USA, 2020; Volume 33, pp. 9422–9434. [Google Scholar]
7. Chan, K.H.R.; Yu, Y.; You, C.; Qi, H.; Wright, J.; Ma, Y. ReduNet: A White-box Deep Network from the Principle of Maximizing Rate Reduction. J. Mach. Learn. Res. 2022, 23, 1–103. [Google Scholar]
8. Li, Z.; Chen, Y.; LeCun, Y.; Sommer, F.T. Neural Manifold Clustering and Embedding. arXiv 2022, arXiv:2201.10000. [Google Scholar] [CrossRef]
9. Canzar, S.; Do, V.H.; Jelić, S.; Laue, S.; Matijević, D.; Prusina, T. Metric Multidimensional Scaling for Large Single-Cell Data Sets using Neural Networks. bioRxiv 2021. [Google Scholar] [CrossRef]
10. Salakhutdinov, R.; Hinton, G. Learning a Nonlinear Embedding by Preserving Class Neighbourhood Structure. In Machine Learning Research, Proceedings of the Eleventh International Conference on Artificial Intelligence and Statistics, San Juan, Puerto Rico, 21–24 March 2007; PMLR: San Juan, Puerto Rico, 2007; Volume 2, pp. 412–419. [Google Scholar]
11. van der Maaten, L. Learning a Parametric Embedding by Preserving Local Structure. In Machine Learning Research, Proceedings of the Twelth International Conference on Artificial Intelligence and Statistics, Hilton Clearwater Beach Resort, Clearwater Beach, FL, USA, 16–18 April 2009; PMLR: Hilton Clearwater Beach Resort, Clearwater Beach, FL, USA, 2009; Volume 5, pp. 384–391. [Google Scholar]
12. Vincent, P.; Larochelle, H.; Lajoie, I.; Bengio, Y.; Manzagol, P.A. Stacked Denoising Autoencoders: Learning Useful Representations in a Deep Network with a Local Denoising Criterion. J. Mach. Learn. Res. 2010, 11, 3371–3408. [Google Scholar]
13. Le, Q.V. Building high-level features using large scale unsupervised learning. In Proceedings of the 2013 IEEE International Conference on Acoustics, Speech and Signal Processing, Vancouver, BC, Canada, 26–31 May 2013; pp. 8595–8598. [Google Scholar] [CrossRef]
14. Hinton, G.E.; Salakhutdinov, R.R. Reducing the dimensionality of data with neural networks. Science 2006, 313, 504–507. [Google Scholar] [CrossRef] [PubMed]
15. Paszke, A.; Gross, S.; Massa, F.; Lerer, A.; Bradbury, J.; Chanan, G.; Killeen, T.; Lin, Z.; Gimelshein, N.; Antiga, L.; et al. PyTorch: An Imperative Style, High-Performance Deep Learning Library. In Advances in Neural Information Processing Systems 32; Curran Associates, Inc.: New York, NY, USA, 2019; pp. 8024–8035. [Google Scholar]
16. Kingma, D.; Ba, J. Adam: A Method for Stochastic Optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
17. Hemberg Group at the Sanger Institute. scRNA Seq Datasets. Available online: https://hemberg-lab.github.io/scRNA.seq.datasets/ (accessed on 3 April 2023).
18. Duò, A.; Robinson, M.D.; Soneson, C. A systematic performance evaluation of clustering methods for single-cell RNA-seq data. F1000Res 2018, 7, 1141. [Google Scholar] [CrossRef] [PubMed]
19. Do, V.H.; Rojas Ringeling, F.; Canzar, S. Linear-time cluster ensembles of large-scale single-cell RNA-seq and multimodal data. Genome Res. 2021, 31, 677–688. [Google Scholar] [CrossRef] [PubMed]
20. Dumitrascu, B.; Villar, S.; Mixon, D.G.; Engelhardt, B.E. Optimal marker gene selection for cell type discrimination in single cell analyses. Nat. Commun. 2021, 12, 1186. [Google Scholar] [CrossRef] [PubMed]
21. Bellman, R. Dynamic Programming; Princeton University Press: Princeton, NJ, USA, 1957. [Google Scholar]
22. Vinyals, O.; Fortunato, M.; Jaitly, N. Pointer Networks. In Advances in Neural Information Processing Systems; Curran Associates, Inc.: New York, NY, USA, 2015; Volume 28. [Google Scholar]
Figure 1. An example of a dumbbell for a pair of sets $( A , B )$.
Figure 1. An example of a dumbbell for a pair of sets $( A , B )$.
Figure 2. In both figures, the x-coordinate stands for the dimension, while the y coordinate stands for the number of dumbbells divided by the size of all pairwise distances, i.e., $n ( n − 1 ) / 2$. (left) Size of the dataset is $n = 5000$. Dumbbells are computed using the WSPD algorithm proposed in , and the size of a WSPD (i.e., the number of dumbbells) is reported. (right) The number of dumbbells stays constant with the NN-WSPD approach even when the dimension grows.
Figure 2. In both figures, the x-coordinate stands for the dimension, while the y coordinate stands for the number of dumbbells divided by the size of all pairwise distances, i.e., $n ( n − 1 ) / 2$. (left) Size of the dataset is $n = 5000$. Dumbbells are computed using the WSPD algorithm proposed in , and the size of a WSPD (i.e., the number of dumbbells) is reported. (right) The number of dumbbells stays constant with the NN-WSPD approach even when the dimension grows.
Figure 3. The boxplots show on the vertical axes the bounds given by Equation (5), i.e., $1 + 4 / s$, for different datasets. Note that for dimension 2 a WSPD is computed directly on the input dataset (leftmost boxplot in each image). For all other dimensions greater than two, autoencoder (AE) and mMDS neural network are used to map the input points to dimension two, and the bound $1 + 4 / s$ is shown for the reconstructed dumbbells in the original dimension. All boxplots on the right show the bound on the reconstructed dumbbells after the refining step. The blue line drawn on the plots in the right column shows the multiplicative increase in dumbbells before and after the refining stage.
Figure 3. The boxplots show on the vertical axes the bounds given by Equation (5), i.e., $1 + 4 / s$, for different datasets. Note that for dimension 2 a WSPD is computed directly on the input dataset (leftmost boxplot in each image). For all other dimensions greater than two, autoencoder (AE) and mMDS neural network are used to map the input points to dimension two, and the bound $1 + 4 / s$ is shown for the reconstructed dumbbells in the original dimension. All boxplots on the right show the bound on the reconstructed dumbbells after the refining step. The blue line drawn on the plots in the right column shows the multiplicative increase in dumbbells before and after the refining stage.
Figure 4. The boxplots show on the vertical axes the bounds given by Equation (5), i.e., 1+4/s, for two real-world scRNA-seq datasets from the . NN-WSPD was applied with autoencoder (AE) and mMDS neural networks. Note that the integer number above the boxplots that show the quality of refined dumbbells states the increase in the number of dumbbells after the refining step of the NN-WSPD algorithm.
Figure 4. The boxplots show on the vertical axes the bounds given by Equation (5), i.e., 1+4/s, for two real-world scRNA-seq datasets from the . NN-WSPD was applied with autoencoder (AE) and mMDS neural networks. Note that the integer number above the boxplots that show the quality of refined dumbbells states the increase in the number of dumbbells after the refining step of the NN-WSPD algorithm.
 Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

## Share and Cite

MDPI and ACS Style

Matijević, D. Well-Separated Pair Decompositions for High-Dimensional Datasets. Algorithms 2023, 16, 254. https://doi.org/10.3390/a16050254

AMA Style

Matijević D. Well-Separated Pair Decompositions for High-Dimensional Datasets. Algorithms. 2023; 16(5):254. https://doi.org/10.3390/a16050254

Chicago/Turabian Style

Matijević, Domagoj. 2023. "Well-Separated Pair Decompositions for High-Dimensional Datasets" Algorithms 16, no. 5: 254. https://doi.org/10.3390/a16050254

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.