1. Introduction
Modeling datasets as metric spaces seems to be natural for some applications and concepts revolving around the Gromov–Hausdorff distance—a notion of distance between compact metric spaces—provide a useful language for expressing properties of data and shape analysis methods. In many situations, however, this is not enough, and one must incorporate other sources of information into the model, with “weights” attached to each point being one of them. This gives rise to the idea of representing data as metric measure spaces, which are metric spaces endowed with a probability measure. In terms of a distance, the Gromov–Hausdorff metric is replaced with the Gromov–Wasserstein metric.
1.1. Notation and Background Concepts
The book by Burago,
et al. [
1] is a valuable source for many concepts in metric geometry. We refer the reader to that book for any concepts not explicitly defined in these notes.
We let denote the collection of all compact metric spaces and by the collection of all isometry classes of . Recall that for a given metric space , its diameter is defined as Similarly, the radius of X is defined as
For a fixed metric space , we let denote the Hausdorff distance between (closed) subsets of Z.
We will often refer to a metric space by only X, but the notation for the underlying metric will be implicitly understood to be . Recall that a map between metric spaces and is an isometric embedding if for all . The map φ is an isometry if it is a surjective isometric embedding.
Recall that given measurable spaces and , a measure μ on and a measurable map , the push-forward measure on acts according to for any .
A metric measure space (mm-space for short) is a triple where is a compact metric space and is a Borel probability measure with full support: . We denote by the collection of all mm-spaces. An isomorphism between is any isometry , such that
2. The Gromov–Hausdorff Distance
One says that a subset is a correspondence between sets X and Y whenever and where and are the canonical projections. Let denote the set of all correspondences between X and Y.
The Gromov–Hausdorff (GH) distance between compact metric spaces
and
is defined as:
where
R ranges over
Example 1. The GH distance between any compact metric space X and the space with exactly one point is equal to
It turns out that is a nice space in that is has many compact subclasses.
Theorem 1. ([
1])
Let be a bounded function and . Let be any family of compact metric spaces, such that for all , such that for any , any admits an ε-net with at most elements. Then, is pre-compact in the Gromov–Hausdorff topology. Example 2. An important example of families, such as the above, is given by those closed n-dimensional Riemannian manifolds with the diameter bounded by and the Ricci curvature bounded below by κ.
Theorem 2 ([
2]).
The space is complete. It then follows from the two theorems above that classes , such as above, are totally bounded for the Gromov–Hausdorff distance. This means that such classes are easy to organize in the sense of clustering or databases.
In many practical applications, one would like to take into account “weights” attached to points in a dataset. For example, the two metric spaces with the weights below are isometric, but not isomorphic in the sense that no isometry respects the weights:
The idea is that weights represent how much we trust a given “measurement” in practical applications. This leads to considering a more general collection of datasets and, in turn, an adapted notion of equality and a compatible metric over them. This naturally leads to regarding datasets as mm-spaces and then finding a notion of distance on compatible with isomorphism of mm-spaces.
3. A Metric on
Let and be two given mm-spaces. In our path to defining a distance between mm-spaces, we emulate the construction of the Gromov–Hausdorff distance and start by identifying a notion of correspondence between mm-spaces.
A probability measure μ over is called a coupling between and if and We denote by the collection of all couplings between and .
Example 3. When , , and thus, there is a unique coupling between X and Y:
Example 4. Consider for example the spaces with two points each that we depicted above. In that case,
can be identified with the vector
and
with the vector
In this case, one sees that the matrix:
induces a valid coupling.
Now, given
, consider the function
and pick any
One then integrates this function against the measure
and infimizes over the choice of
to define the Gromov–Wasserstein distance of order
p [
3]:
Remark 1. This is an
analogue of Equation (
1).
Theorem 3 ([
3]).
The Gromov–Wasserstein distance of order defines a proper distance on the collection of isomorphism classes of mm-spaces. By standard compactness arguments, one can prove that the infimum above is always attained [
3]. Let
denote the set of all the couplings in
that achieve the minimum. The structure of the former set depends not only on
and
, but also on
,
and
p.
Example 5. Consider the mm-space with exactly one point:
. Then,
and we define
—the
p-statistical diameter of
X—as twice the right-hand side. Notice that
is equal to the usual diameter of
X (as a metric space).
Question 1. To what extent are we able to replicate the nice properties of in the context of ? In particular, it is of interest to investigate whether this new space of datasets is complete and whether one can easily identify rich pre-compact classes.
3.1. Pre-Compactness
Theorem 4 ([
3]).
For a function non-decreasing , such that for and , let denote the set of all mm-spaces X, such that and for all . Then, is pre-compact for the Gromov–Wasserstein topology, for any . Remark 2. Recall Example 2, where closed
n-dimensional Riemannian manifolds were regarded as metric spaces. One can, all the same, embed closed Riemannian manifolds into
via
, where
is the geodesic distance induced by the metric tensor
and
stands for the normalized volume measure on
X. It is well known [
4] that for
small,
, where
is the scalar curvature of
X at
x, and
is the total volume of
X. Thus, a lower bound on
plays the role of a proxy for an upper bound on curvature.
3.2. Completeness
The Space with any p-Gromov–Wasserstein distance is not complete. Indeed, consider the following family of mm-spaces: , where consists of points at distance one from each other, and all with weights .
Claim 1. For all , .
The claim will follow from the following claim and triangle inequality for :
Claim 2. For all ,
In order to verify the claim, we denote by
the points of
and label the points in
by
Consider the following coupling between
and
, the reference measures on
and
:
It is clear that this defines a valid coupling between and
Now, by definition, , so the claim follows.
Claim 1 indicates that constitutes a Cauchy sequence in . However, a potential limit object for this sequence will have countably infinitely many points at distance one from each other. This space is not compact, thus is not a complete metric.
3.3. Other Properties: Geodesics and Alexandrov Curvature
Very recently, Sturm [
5] pointed out that
is a geodesic space when endowed with any
,
. This means that given any two spaces
in
, one can find a curve
, such that
Proposition 1 ([
5]).
For each , the space is geodesic. Furthermore, for , the following curves on define geodesics between and in :where for and Furthermore, for , all geodesics are of this form. Sturm further proved that the completion of the space with metric satisfies:
Theorem 5 ([
5]).
The metric space is an Alexandrov space of curvature . Amongst the consequences of this property is the fact that one can conceive of gradient flows on the space of all mm-spaces [
5].
3.4. The Metric in Applications
Applications of the notion of Gromov-Wasserstein distance arise in shape and data analysis. In shape analysis, the main application is shape matching under invariances. Many easily computable lower bounds for the GW distance have been discussed in [
3,
6]. All of them lead to solving linear programming optimization problems (for which there are polynomial time algorithms) or can be computed via elicit formulas. As an example, consider the following invariant of an mm-space
:
This invariant simply encodes the distribution of pairwise distances on the dataset
X, and it is defined by analogy with the so-called shape distributions that are well known in computer graphics [
7]. Then, one has:
Proposition 2 ([
3,
6]).
Let be any two mm-spaces and . Then, Remark 3. This invariant is also related to the work of Boutin and Kemper [
8] and Brinkman and Olver [
9].
Other lower bounds which can be computed in time polynomial in the number of points in the underlying mm-spaces have been reported in [
3]. As a primary example, one has that the
local shape distributions of shapes provide a lower bound which is strictly stronger than the ones in the Proposition above. In more detail, consider for a given mm-space
the invariant:
Then, for mm-spaces
X and
Y consider the
cost function
given by:
One then has:
Proposition 3 ([
3,
6]).
Let be any two mm-spaces and . Then,where μ ranges in Remark 4. Solving for the infimum above leads to a mass transportation problem for which there exists efficient linear programming techniques.
Remark 5. It is possible to define a notion of
spectral Gromov-Wasserstein distance which operates at the level of compact Riemannian manifolds without boundaries, and is based con the comparison of heat-kernels. This notion permits inter-relating many pre-existing shape matching methods and suggests some others [
12].
4. Discussion and Outlook
The Gromov–Hausdorff distance offers a useful language for expressing different tasks in shape and data analysis. Its origins are in the work of Gromov on synthetic geometry. For finite metric spaces, the Gromov–Hausdorff distance leads to solving NP-hard combinatorial optimization problems. Related to construction is Gromov–Wasserstein distances that operate on metric measure spaces [
3,
10]. In contrast to the Gromov–Hausdorff distance, the computation of Gromov–Wasserstein distances leads to solving quadratic optimization problems on continuous variables. The space of all metric measures spaces endowed with a certain variant of the Gromov–Wasserstein distance [
3] enjoys nice theoretical properties [
5]. It seems of interest to develop provably correct approximations to these distances when restricted to some suitable subclasses of finite metric spaces. Other aspects of the Gromov–Wasserstein distance are discussed in [
3,
5,
10,
11,
12].