Abstract
This paper examines the Wasserstein metric between the empirical probability measure of n discrete random variables and a continuous uniform measure in the d-dimensional ball, providing an asymptotic estimation of their expectations as n approaches infinity. Furthermore, we investigate this problem within a mixed process framework, where n discrete random variables are generated by the Poisson process.
MSC:
60B10; 60G57
1. Introduction
Article [1] investigates the Ollivier curvature of random geometric graphs, with a key step being the estimation of Wasserstein metrics between the empirical probability measure of n discrete random variables and a continuous uniform one in a d-dimensional ball. The authors applied results from [2], which are based on the interval , whereas Ollivier curvature is built in balls. To address this discrepancy, we aim to refine the proof based on balls in order to enhance the robustness and accuracy of the process described in [1] and to make it suitable for our purposes. Furthermore, since [2] requires , we extended our upper and lower bounds estimation to include the case , aligning our research objectives with the broader scope of the study.
Additionally, lattice methods used in statistical mechanical approaches [3] often involve similar notations and convergence from discrete physical quantities to continuous ones, suggesting potential connections with convergences from discrete probabilities to continuous ones. For instance, consider a collection of point charges denoted as and their corresponding locations represented by the independent and uniformly distributed random variables , where represents a bounded region within three-dimensional space with its volume defined as . Assuming an ideal scenario in implicit solvation models for biological molecules, it can be postulated that each charge possesses the value of , thereby establishing a discrete charge density expressed as . On the other hand, we consider a continuum charge density represented by a uniform measure in an ideal scenario. Thus, transitioning from discrete (i.e., point) charges to a continuum charge density can be a pathway from a discrete probability measure to a continuous one in Wasserstein metrics. Consequently, we may contemplate convergences of the corresponding discrete electrostatic energies and others in terms of Wasserstein metrics or within Wasserstein spaces.
The authors in [4] have provided estimations for the convergence rate in Wasserstein metrics of empirical measures on complex computational cases, involving numerous asymptotic calculations. Our findings are consistent with their results in corresponding scenarios. In comparison, our proof primarily relies on estimating the expectations of optimal matching problems to obtain upper bounds on the expectations of Wasserstein metrics. We chose this technique because, as mentioned in [5], (i) the definition of Wasserstein metrics makes them convenient for problems involving optimal transports, such as those arising from partial differential equations; (ii) Wasserstein metrics possess a rich duality property which is particularly useful when considering (2) (in contrast to bounded Lipschitz distances), and passing back and forth from the original to the dual definition is often technically convenient; (iii) being defined by an infimum, it is often relatively straightforward to bound Wasserstein metrics from above by constructing couplings between and ; and (iv) Wasserstein metrics incorporate a lot of the geometry of the space. For instance, the mapping is an isometric embedding of into (Wasserstein space of order p), but there are much deeper links. This partly explains why a Wasserstein space of is often very well adapted to statements that combine weak convergence and geometry.
Motivated by the virtues of Wasserstein spaces and these inspirations, we aim to bridge the gap between discrete probabilities and their continuous counterparts using Wasserstein metrics as measures in our study. Recently, significant advancements have been made in the research progress concerning the rate of convergence of Wasserstein metrics. For instance, in [6], the authors investigated the precise rate of convergence of the quadratic Wasserstein metric between empirical measures and uniform distributions on by employing well-known techniques from partial differential equations. Additionally, in [7], researchers explored upper bounds for the mean Wasserstein metric between two probabilities on where , using Fourier transformation, and subsequently applied these findings to estimate the mean Wasserstein metric between two empirical measures under certain assumptions. Furthermore, in [8], an author examined upper bounds for the mean rate within the quadratic Wasserstein metric on a d-dimensional compact Riemannian manifold where . Notably, there are also ongoing studies focusing on higher-order (p-th order) Wasserstein metrics; however, we refrain from listing them here.
2. Preliminary Estimation
Definition 1.
Let be independent and uniformly distributed random variables in a d-dimensional ball where represents the Euclidean metric in The random variable represents the optimal matching between and with σ iterating over all permutations of
By applying the dual principle [9,10], or referring to the proof process of Lemma 1 in [2], we have
where the set of Lipschitz functions It is worth noting that every Lipschitz function in can be extended to one in
Therefore, we have The following Lemma 1 gives an upper and lower bound estimation for the expectation
Lemma 1
(Optimal matching). For the above optimal matching problem, we have
in dimension and
in dimension
Proof.
We provide a detailed proof, referring to Equation (A1) in Appendix A and Equation (A3) in Appendix B. The method employed here is essentially based on the work of [2], with several improvements and modifications made to extend its applicability to random variables within balls.
3. Main Results and Proofs
The following results present Wasserstein metrics between empirical and uniform measures in d-dimensional balls. Generally, a Wasserstein metric between two probability measures is defined as follows:
Definition 2.
Let and be Borel probability measures in a compact metric space and let denote the set of all couplings of and , i.e.,
A Wasserstein metric is defined as
By applying the duality principle (Kantorovich Dual Theorem) in Chapter 6, Remark 6.5 of [5], we can express the Wasserstein metric as
where denotes the set of Lipschitz functions based on the metric of with a coefficient of 1. From the duality formula, we can further assume that any function satisfying
Notice: In all subsequent discussions, we will explicitly specify that the metric being considered is a Euclidean metric. Additionally, we will adopt the notation for a sequence , where is a constant. This notation implies the existence of positive constants C such that as n is large enough.
Theorem 1.
Let be independent and uniformly distributed random variables in a d-dimensional ball The empirical measure is given by
which represents the proportion of points in the sample that lie in a measurable subset A of . demotes the uniform measure in As the sample size n tends to infinity, it can be shown that an expected Wasserstein metric between and , denoted as , decays at a rate
Proof.
Now, we consider a Wasserstein metric in with and then
Let be independent uniformly distributed random variables in and then
So
where and hence from Lemma 1, it has
□
Next, we consider an empirical measure and a uniform measure in a ball
Corollary 1.
In general, let be independent and random variables uniformly distributed in the d-dimensional ball with radius Consider an empirical measure and a uniform measure in , where represents the empirical measure defined as
for a measurable subset A of , and denotes the uniform measure in Then, it follows that
Proof.
Consider the map , defined by where Thus, and correspond to the empirical measure and the uniform measure in , respectively, which establishes a one-to-one relationship between empirical measures in and those in , as well as between uniform measures in and those in . In particular, we can write
Therefore, from Theorem 1 we obtain
□
We next generalize Theorem 1 to the case where the number of random variables, denoted by N, follows a Poisson distribution with a parameter and is independent of these random variables. This case actually corresponds to a specific spacial Poisson process in [1] with intensity measure which describes a spatial configuration of points in the ball Here, denotes the volume measure in d-dimensional Euclidean space. Moreover, in [1], it is also stated that representing the number of random points in , called size, and the parameter of the Poisson distribution is equivalent to , derived from the corresponding Poisson point process. The notation represents slight perturbations of n in order to observe how a expected Wasserstein metric gradually changes as n approaches infinity.
Theorem 2.
Let denote the empirical random measure with respect to independent and uniformly distributed random variables in , defined as
for a measurable subset A of . N follows a Poisson distribution with a parameter , which is independent of random variables . Let represent the uniform measure in the aforementioned ball. Then, it follows that
Proof.
Since the number N follows a Poisson distribution with the mean and are uniformly distributed random variables in , which are independent of N, we have
According to Theorem 1, it follows that
On the other hand, from Lemma 1.2 in [11], one can obtain
where is a constant, and
Let us denote
Then, we obtain an expression for the expected value as follows:
We further estimate these three terms and find that
and
For term it is bounded as follows:
Since c was arbitrary, with a suitable adjustment in constant c, we conclude that
□
Now, we consider a d-dimensional ball . The number of random variables in , still denoted as N, follows a Poisson distribution with a parameter , and N is independent of these random variables. This actually corresponds to a spatial Poisson process with intensity measure in , as discussed in [1], and the parameter of Poisson distribution is equivalent to , derived from the corresponding Poisson point process.
Corollary 2.
Let and . We denote by the empirical measure with respect to independent and uniformly distributed random variables in , i.e.,
for a measurable subset A of . N follows a Poisson distribution with a parameter and is independent of random variables . Let be the uniform measure in Then, we have
Proof.
Combining the proof in Theorem 2 and Corollary 1, we first note that N follows a Poisson distribution in with mean value . Therefore, we can obtain
□
4. Conclusions
The result in Corollary 2 can be directly applied to produce Appendix A.3 in [1]. We have successfully refined the proof based on balls, thereby enhancing the robustness and accuracy of the process described in [1]. Furthermore, our study has effectively bridged the gap between discrete probabilities and their continuous counterparts by utilizing Wasserstein metrics as approach measures. Moving forward, we aim to apply our methodology to analyze lattice problems in statistical mechanical approaches that involve similar notation and convergence from discrete physical quantities to continuous ones, such as electrostatic approach problems.
We derived the upper bound for the convergence rate of the Wasserstein distance between a uniform distribution and its empirical distribution when using the dual method. Our result is consistent with the order of convergence rate in [4], but we provide a specific constant term. Furthermore, we extended this analysis to estimate the convergence rate of random multinomial empirical distributions towards uniform distributions, yielding similar results. However, our approach does not apply to the case when , and we did not obtain a lower bound estimation for the convergence rate. In real-world scenarios, connections between discrete and continuous worlds can be established through random graphs by extending mathematical concepts from manifolds to graphs. For instance, in [1], authors generalize the Ollivier graph curvature definition to enhance its versatility and prove that the Ollivier curvature of random geometric graphs in Riemannian manifolds converges to the Ricci curvature of the manifold. Additionally, Appendix C3 in [1] also provides methods for computing Wasserstein metrics through simulations.
Author Contributions
Conceptualization, W.Y., X.Z. and X.W.; methodology, W.Y. and X.Z.; validation, W.Y., X.W. and X.Z.; formal analysis, W.Y., X.W. and X.Z.; writing—original draft preparation, W.Y.; writing—review and editing, W.Y. and X.Z.; project administration, W.Y. All authors have read and agreed to the published version of the manuscript.
Funding
This research was funded by a school–enterprise cooperation project: Application of hyperbolic network model in data analysis.
Data Availability Statement
No new data were created or analyzed in this study. Data sharing is not applicable to this article.
Conflicts of Interest
The authors declare no conflict of interest. The funder had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.
Appendix A. The Lower Bound of
Since
it follows that
Let the set of points and then
and
Thus, we have
Finally, by Fatou’s lemma, one has
Appendix B. The Upper Bound of
Let so that as , and where Define
Define
, and then we have if First decompose , as follows:
so one has the inequality
Further, we will estimate two parts of in Appendix B.1 and in Appendix B.2, respectively. Combining them yields the following bound for the expectation of :
Appendix B.1. The Estimation of
Since f is Lipschitz, we have
It should be noted that a more optimized estimation for can be found in [2].
Appendix B.2. The Estimation of
Decompose as follows:
We will estimate these three parts separately in the following smaller subsections to obtain
Appendix B.2.1. The Estimation of
According to and the value (A2) of in , we have
Consequently, we can obtain
Appendix B.2.2. The Estimations of and
Estimating this part is challenging, and one may employ convolution decomposition to impose f in small areas. Consequently, the following estimation holds:
Initially, we assume that f represents an indicator function for a set where A is a measurable subset of and estimate Thus, we have
By considering different cases of and we obtain the inequality
Furthermore, let us set By using Formula (A4), we obtain
Finally, we decompose f into the sum of some well-defined convolutions to estimate these components. Since a Lipschitz function f in with can be extended to a Lipschitz function in the entire space with , we consider the function defined as (1). We then decompose it as follows, where ⋯,
and q denoted by Therefore, we have
For the first expectation mentioned above, we have
since and
For the expectation about with since , we have
and on the other hand, we have
For the last expectation about , the above argument still works. Since , we have
If one has If we may obtain , and hence
References
- van der Hoorn, P.; Lippner, G.; Trugenberger, C.; Krioukov, D. Ollivier-Ricci curvature convergence in random geometric graphs. Phys. Rev. Res. 2021, 3, 013211. [Google Scholar] [CrossRef]
- Talagrand, M. Matching random samples in many dimensions. Ann. Appl. Probab. 1992, 2, 846–856. [Google Scholar] [CrossRef]
- Kralj-Iglič, V.; Iglič, A. A Simple Statistical Mechanical Approach to the free Energy of the Electric Double Layer Including the Excluded Volume Effect. J. Phys. II 1996, 6, 477–491. [Google Scholar] [CrossRef]
- Fournier, N.; Guillin, A. On the rate of convergence in Wasserstein distance of the empirical measure. Probab. Theory Relat. Fields 2015, 162, 707–738. [Google Scholar] [CrossRef]
- Villani, C. Optimal Transport: Old and New Part 1; Grundlehren der Mathematischen Wissenschaften, 338. A Series of Comprehensive Studies in Mathematics; Springer: Berlin/Heidelberg, Germany, 2009. [Google Scholar]
- Ambrosio, L.; Stra, F.; Trevisan, D. A PDE approach to a 2-dimensional matching problem. Probab. Theory Relat. Fields 2019, 173, 433–477. [Google Scholar] [CrossRef]
- Bobkov, S.G.; Ledoux, M. A simple Fourier analytic proof of the AKT optimal matching theorem. Ann. Appl. Probab. 2021, 31, 2567–2584. [Google Scholar] [CrossRef]
- Borda, B. Empirical measures and random walks on compact spaces in the quadratic Wasserstein metric. Ann. Inst. Henri Poincaré Probab. Stat. 2023, 59, 2017–2035. [Google Scholar] [CrossRef]
- Kuhn, H.W. The Hungarian method for the assignment problem. Nav. Res. Logist. Q. 1955, 2, 83–97. [Google Scholar] [CrossRef]
- Papadimitriou, C.H.; Steiglitz, K. Combinatorial Optimization, Algorithms and Complexity; Prentice-Hall: Englewood Cliffs, NJ, USA, 1982. [Google Scholar]
- Penrose, M. Random Geometric Graphs; Oxford University Press: Oxford, UK, 2003. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).