Significance
Given a set of points on the sphere, an allocation is a way to divide the sphere into cells of equal area, each associated with a point of . Given two sets of points and on the sphere, a matching is a bijective map from to . Allocation and matching rules that minimize the distance between matched points are related to optimal transport and discretization of continuous measures. We define a matching and allocation rule by considering the gravitational field associated with the point configurations and show that they are optimal in expectation up to multiplication by a constant when our points are chosen independently and uniformly at random.
Keywords: bipartite matching, allocation, transportation, gravity
Abstract
Given a collection of points on a sphere of surface area , a fair allocation is a partition of the sphere into parts each of area 1, and each is associated with a distinct point of . We show that, if the points are chosen uniformly at random and if the partition is defined by a certain “gravitational” potential, then the expected distance between a point on the sphere and the associated point of is . We use our result to define a matching between two collections of independent and uniform points on the sphere and prove that the expected distance between a pair of matched points is , which is optimal by a result of Ajtai, Komlós, and Tusnády.
Suppose that we are given points on the unit sphere . We would like to partition the sphere into equally sized cells, assigning each point to a different cell. How can we make this partition so that each point is close to the points in the cell to which it has been assigned? This natural question, known as the fair allocation problem, has connections to optimal transport and discretization (or “quantization”) of continuous measures (1, 2). Allocation is also closely related to the matching problem, in which red points and blue points are chosen from the sphere (say, independently at random), and our goal is to pair each red point with a different blue point so as to make the distances between paired points as small as possible. Minimal matching for random points in the plane has generated a substantial literature in its own right (3–5).*
We analyze a particular allocation rule called gravitational allocation and apply it to matchings. Gravitational allocation is based on treating our points as wells of a potential function. The cell allocated to a given point is then taken to be the basin of attraction of with respect to the flow induced by the negative gradient of this potential. When the potential takes a particular form that mimics the gravitational potential of Newtonian mechanics, it is ensured that each cell has area 1 (Fig. 1).
Fig. 1.
Gravitational allocation to uniform and independent points on a sphere with and 750. The basin of attraction of each point has equal area. The basins become more elongated as grows, reflecting Theorem 1. The MATLAB script used to generate the gravitational allocation figures in this article is based on code written by Manjunath Krishnapur.
Related Work
The idea of transportation between measures via gradient flows dates back at least to Dacorogna and Moser (6). However, the first analysis that we know of concerning the resulting allocation cells was carried out by Nazarov, Sodin, and Volberg (7), who studied allocation to zeroes of a Gaussian analytic function.
The term gravitational allocation was introduced by Chatterjee, Peled, Peres, and Romik (8, 9), who studied fair allocations to a Poisson process in with . In that setting, they proved exponential tail bounds on the diameter of a typical cell, showing that this diameter is of constant order.
The same does not hold when : it was shown in refs. 10 and 11 that, for translation invariant allocation schemes in or , the expected allocation distance must be infinite. For this reason, to understand what happens when , it helps to consider a finite setting, such as the sphere. Suppose that we take the scaling where each cell has unit area. Then, it turns out that the typical allocation distance will need to be of at least order , which is also the same asymptotic behavior that is seen in minimal matching (3). In a recent paper, Ambrosio, Stra, and Trevisan (12) proved a more precise estimate of for the expectation of the minimum average squared distance between random points and the uniform measure, confirming a prediction of Caracciolo, Lucibello, Parisi, and Sicuro (13).
Other than gravitational allocation, other allocation schemes have been proposed and analyzed, many based on the Gale–Shapley stable matching algorithm (11, 14–16.
Formal Definitions and Main Result
Let denote the sphere centered at the origin with surface area , so that we work in the scaling where each cell has unit area. Let denote the surface area measure on , so that .
For any set consisting of points, we say that a measurable function is a fair allocation of to if it satisfies the following:
| [1] |
For , we call the cell allocated to .
Let us now describe gravitational allocation in particular. First, we define a potential function given by
| [2] |
where denotes Euclidean distance in . For each location , let denote the negative gradient of with respect to the usual spherical metric (i.e., the one induced from ). We can view as lying in the plane tangent to at (i.e., the tangent space), so that is a vector field on .
Second, we consider the flow induced by . For any , let denote the integral curve that solves the differential equation
| [3] |
By standard results about ordinary differential equations, the curve can be defined up until some maximal time (possibly ). In fact, will be finite for all ,† because by flowing along , will eventually fall into one of the wells of the potential (i.e., one of the points in ) (Fig. 2).
Fig. 2.
Illustration of , , and for and .
We thus define the basin of attraction of as
| [4] |
(i.e., the set of points that will eventually flow into ). We then define the gravitational allocation function to be
| [5] |
It turns out that indeed defines a fair allocation of to , so that each has area 1. Before explaining why this is the case, let us first state our main result.
Theorem 1.
Let be a positive integer. Consider any , and let be a set of points chosen uniformly and independently at random from . Then, there is a constant such that
[6] More generally, for any , there is a constant depending only on such that
[7]
Why Is Gravitational Allocation a Fair Allocation?
The reader may find it somewhat surprising that the basins of attraction in gravitational allocation always have equal areas, even if a point in is crowded by many other points in (Fig. 3). As seen in Fig. 3, the surrounded point will still attract certain faraway points, so that its basin of attraction still has total area 1.
Fig. 3.
The center point is surrounded by seven other nearby points (Left). Nevertheless, it turns out that its basin of attraction (light blue; Right) can slip past its neighbors in certain places.
We give two explanations for this phenomenon. Both explanations rely on the fact that our potential satisfies the Poisson equation
where denotes the spherical Laplacian (i.e., the Laplace–Beltrami operator on ).
The first explanation is based on the divergence theorem. Consider any and its cell . Since is a basin of attraction, must be parallel to along its boundary. We can then apply the divergence theorem‡to obtain
It follows that as desired.
The second explanation is slightly longer, but it also provides a more detailed understanding of the flow under . Imagine the surface area measure as representing the density of grains of sand uniformly distributed on the sphere. The sand is flowing along , so that a grain of sand at will be moved to location after time .
In a small time , the net change in the density of sand at a point will be approximately
Thus, the density is decreasing everywhere at a uniform rate, except at points of , where sand is accumulating (at the same rate for each point). Integrating this over time, the density of sand at a time will be given by
We find that , so that the amount of sand at each point in tends to one. Consequently, the area of each basin of attraction must have been one.
Proof Outline of the Main Theorem
The proof of Theorem 1 is based on estimating the magnitude of the gradient force . In the previous section, we saw that, after time , all but a proportion of the sphere will have reached one of the points in , and therefore, the average time that it takes for a point to flow into a potential well is . We can also estimate the average distance traveled in a similar way:
| [8] |
It remains to estimate the average magnitude of , which is given by the following lemma.
Lemma 2. Fix any . Then, , where the expectation is taken over the randomness of .
Taking expectations in Eq. 8 and then integrating Lemma 2 over all proves Theorem 1 in the case . Larger values of can be handled in the same spirit, but it requires more involved estimates for that we do not reproduce here (ref. 17 has details).
Proof: Let and , so that and . Thus, represents the contribution to coming from the point .
To estimate , it is convenient to decompose into the contributions of nearby and faraway points in . For our purposes, “near” means points within the spherical cap of radius 1 around , which we denote by . Then, we may write
| [9] |
When , an explicit computation shows that is of order . It is also not hard to calculate that the expected number of points in with distance from that is between and is of order . By the triangle inequality, we can estimate as
| [10] |
To estimate the far term, the triangle inequality is too weak, because we expect much cancellation between the . In fact, by symmetry, we have . Thus, we instead estimate the second moment
| [11] |
which is the bound claimed in Lemma 2.
A Heuristic Picture
Lemma 2 also provides a good heuristic proof of Eq. 7. We know by Lemma 2 that, for a typical point , we have , and moreover, our above analysis suggests that the value of is dominated by contributions from faraway points. Thus, we expect that direction and speed of travel for under the flow induced by will remain relatively constant.
However, will not travel forever in this way; suppose that it passes within distance of a point . Then, the contribution from to the overall “force” will be of order , which may overpower the contribution from all other points, causing to fall into the potential well at .
Consider a strip of width around the path of (Fig. 4). If there is a point in this strip, then it is likely to “swallow” (i.e., will be allocated to ). The probability that any given region contains no points of decays exponentially in its area, which suggests the heuristic
giving Eq. 7, because is bounded above by the distance traveled by .
Fig. 4.
The speed is mainly determined by points far away and is approximately constant in large regions, except very near points of . A typical point, therefore, travels in an approximately straight line until it gets within distance of some point in .
From Allocations to Matchings
We now turn to the connection between fair allocations and optimal matchings. Suppose that and are two sets of points in . A matching from to is a bijective function . Recall that the matching problem is to find the matching that minimizes the total distance between matched points.
When the points of and are drawn uniformly at random, the asymptotic behavior of the minimal matching distance was identified by Ajtai, Komlós, and Tusnády (3), who proved the following theorem.
Theorem 3 (Ajtai–Komlós–Tusnády). Suppose that and each consist of points drawn uniformly and independently at random from . Let
Then, there are constants for which
| [12] |
It turns out that the average displacement of a fair allocation gives an upper bound on the matching distance, as the next proposition shows.
Proposition 4. Let be two sets of points, and let and be fair allocations of to and , respectively. Then, there exists a matching such that
| [13] |
Remark 5: Consider the case where and are drawn uniformly at random, and suppose that we use gravitational allocation for and in Proposition 4. Then, the case of Theorem 1 implies that the right-hand side of Eq. 13 has expectation of order . Comparing with Theorem 3, this implies that the asymptotic rate of in Theorem 1 is the best possible up to a constant factor. By Eq. 8, we also get that is at least of order for any fixed .
The triangle inequality for the linear Wasserstein distance justifies why we can pass from an allocation to a matching, but we choose to describe the connection explicitly. Let denote the cell allocated to , and similarly, let . Consider the matrix given by
We see that is a doubly stochastic matrix:
By the Birkhoff–von Neumann theorem (ref. 18, theorem 5.5), any doubly stochastic matrix is a convex combination of permutation matrices. For a permutation , we write to denote the corresponding permutation matrix, so that if and otherwise. Then, we may write
| [14] |
where are nonnegative numbers summing to one and are permutations.
Let be chosen uniformly at random from . Observe that and that on the event . By Eq. 14 and this observation,
| [15] |
By the triangle inequality, the right side of Eq. 15 is bounded above by the right side of Eq. 13, which implies Proposition 4.
Online Matching
One can also consider an “online” version of the matching problem, in which we initially see only the points in , and we are given the points in one by one. As soon as is revealed to us, we must immediately match it to a point in (that has not already been matched). In particular, we make this decision without knowing the locations of the remaining points in .
There is a natural online matching algorithm using gravitational allocation. When a point is revealed, let be the set of points in that have not yet been matched. We then consider the gravitational allocation to and match to .
The analysis of this procedure is particularly simple if the points of and are sampled uniformly and independently at random. Consider what happens when we pair the first point . According to Theorem 1, the expected distance between and its pair is bounded by
Since gives a fair allocation and the first point is drawn uniformly at random, each of the points in is an equally likely match for under our scheme. It follows that the remaining points will still be distributed uniformly and independently at random. Thus, we have reduced the problem to matching two sets of independent random points on after incurring a cost of for matching the first pair (Fig. 5).
Fig. 5.
Illustration of the online matching algorithm. The set consists of uniform and independent points on the sphere of area .
We may iterate this analysis for each point in . When we receive , there will be remaining unpaired points in (still uniformly distributed), so that a typical distance in gravitational allocation will be , where the factor comes from rescaling to . Thus,
which shows that, even in the online setting, one has similar asymptotics as in Theorem 3.
We remark that our online matching algorithm can be implemented efficiently using the well-known “fast multipole method” introduced by Rokhlin (19) and Greengard and Rokhlin (20). This entails precomputing estimates of the gravitational potential from clusters of points in , and these computations can be reused as new points of are introduced.
Gravitational Allocation for Other Point Processes
So far, we have focused on the setting where our points on are taken independently at random. However, one may also analyze other random point processes where the points are not independent, which allows them to be distributed more evenly over the sphere.
One example is given by the roots of a certain Gaussian random polynomial. Specifically, we look at the polynomial
where are independent standard complex Gaussians. The roots of are then random points in the complex plane, which we can bring to the sphere via stereographic projection. More explicitly, let . The function
maps the horizontal plane in to . Then, viewing the as lying in the horizontal plane,
is a rotationally equivariant random set of points on . [The rotational equivariance comes from the particular choice of coefficients for (ref. 21, chapter 2.3).]
Heuristically, the points of are distributed more evenly than independent uniformly random points, because roots of random polynomials tend to “repel” each other (Fig. 6). This can be quantified as follows. Let be the gravitational allocation. Then, we claim that
| [16] |
Fig. 6.
A simulation of gravitational allocation to the zeroes of a Gaussian random polynomial. The cells are evenly proportioned, in contrast with the more elongated shapes seen in Fig. 1.
To prove this, by Eq. 8 and rotational symmetry, it suffices to show that for any point . It is convenient to pick . Then, in the notation of the Proof of Lemma 2, we may calculate that
where we interpret the complex number on the right-hand side as a 2D vector. Thus, we have
which gives a simple expression for in terms of two independent complex Gaussians. Taking expectations of the magnitude, we obtain
which establishes Eq. 16.
Open Problems
We conclude by describing two other matching algorithms for which we do not know a precise analysis.
First, one may consider a dynamic electrostatic version of gravitational allocation. Suppose that the points in () are positive (negative) and that points of different (similar) kinds attract (repulse) each other. After some time, it seems that each point in will collide with a point in , forming a matching. What will be the average distance between the original positions of matched pairs?
Second, in the online matching problem, instead of matching each new point to a point in according to gravitational allocation, suppose that we simply match to the closest point in that has not been matched already. Alternatively, we can reveal and simultaneously and iteratively match closest pairs of points. In other words, we choose such that is minimized, we define , and we repeat with the sets and . What will be the average matching distance in these settings? In the latter setting, ref. 16, theorem 6 suggests an upper bound for the matching distance of . Can this be improved?
Acknowledgments
We thank Manjunath Krishnapur for useful discussions as well as for sharing his code for producing simulations. Most of this work was carried out while N.H. and A.Z. were visiting Microsoft Research; we thank Microsoft Research for the hospitality.
Footnotes
The authors declare no conflict of interest.
See Profile on page 9646.
*We note that many results are stated for points in a square or a 2D torus instead of the sphere. As , all of these settings are essentially equivalent. For the sake of consistency, in this article, we will state everything in terms of the sphere.
†Except for a set of measure zero.
Assuming various smoothness properties, which we do not justify here.
References
- 1.Huesmann M, Sturm KT. Optimal transport from Lebesgue to Poisson. Ann Probab. 2013;41:2426–2478. [Google Scholar]
- 2.Dereich S, Scheutzow M, Schottstedt R. Constructive quantization: Approximation by empirical measures. Ann Inst Henri Poincaré Probab Stat. 2013;49:1183–1203. [Google Scholar]
- 3.Ajtai M, Komlós J, Tusnády G. On optimal matchings. Combinatorica. 1984;4:259–264. [Google Scholar]
- 4.Leighton T, Shor P. Tight bounds for minimax grid matching with applications to the average case analysis of algorithms. Combinatorica. 1989;9:161–187. [Google Scholar]
- 5.Talagrand M. Matching theorems and empirical discrepancy computations using majorizing measures. J Am Math Soc. 1994;7:455–537. [Google Scholar]
- 6.Dacorogna B, Moser J. On a partial differential equation involving the Jacobian determinant. Ann de l’Institut Henri Poincaré Analyse Non Linéaire. 1990;7:1–26. [Google Scholar]
- 7.Nazarov F, Sodin M, Volberg A. Transportation to random zeroes by the gradient flow. Geom Funct Anal. 2007;17:887–935. [Google Scholar]
- 8.Chatterjee S, Peled R, Peres Y, Romik D. Gravitational allocation to Poisson points. Ann Math. 2010;172:617–671. [Google Scholar]
- 9.Chatterjee S, Peled R, Peres Y, Romik D. Phase transitions in gravitational allocation. Geom Funct Anal. 2010;20:870–917. [Google Scholar]
- 10.Holroyd AE, Liggett TM. How to find an extra head: Optimal random shifts of Bernoulli and Poisson random fields. Ann Probab. 2001;29:1405–1425. [Google Scholar]
- 11.Hoffman C, Holroyd AE, Peres Y. A stable marriage of Poisson and Lebesgue. Ann Probab. 2006;34:1241–1272. [Google Scholar]
- 12.Ambrosio L, Stra F, Trevisan D. 2018 A PDE approach to a 2-dimensional matching problem. Probab Theory Relat Fields. Available at https://link.springer.com/article/10.1007%2Fs00440-018-0837-x#citeas. Accessed September 4, 2018.
- 13.Caracciolo S, Lucibello C, Parisi G, Sicuro G. Scaling hypothesis for the Euclidean bipartite matching problem. Phys Rev E. 2014;90:012118. doi: 10.1103/PhysRevE.90.012118. [DOI] [PubMed] [Google Scholar]
- 14.Holroyd AE, Peres Y. Extra heads and invariant allocations. Ann Probab. 2005;33:31–52. [Google Scholar]
- 15.Hoffman C, Holroyd AE, Peres Y. Tail bounds for the stable marriage of Poisson and Lebesgue. Canad J Math. 2009;61:1279–1299. [Google Scholar]
- 16.Holroyd AE, Pemantle R, Peres Y, Schramm O. Poisson matching. Ann Inst Henri Poincaré Probab Stat. 2009;45:266–287. [Google Scholar]
- 17.Holden N, Peres Y, Zhai A. 2017. Gravitational allocation for uniform points on the sphere. arXiv:170408238. Preprint, posted April 26, 2017.
- 18.van Lint JH, Wilson RM. A Course in Combinatorics. 2nd Ed Cambridge Univ Press; Cambridge, UK: 2001. [Google Scholar]
- 19.Rokhlin V. Rapid solution of integral equations of classical potential theory. J Comput Phys. 1985;60:187–207. [Google Scholar]
- 20.Greengard L, Rokhlin V. A fast algorithm for particle simulations. J Comput Phys. 1987;73:325–348. [Google Scholar]
- 21.Hough JB, Krishnapur M, Peres Y, Virág B. 2009. Zeros of Gaussian Analytic Functions and Determinantal Point Processes, University Lecture Series (American Mathematical Society, Providence, RI), Vol 51. [Google Scholar]






