Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2020 Dec 22.
Published in final edited form as: Comput Graph X. 2019 Jun 6;2:100005. doi: 10.1016/j.cagx.2019.100005

A Kernel for Multi-Parameter Persistent Homology

René Corbet a, Ulderico Fugacci a, Michael Kerber a, Claudia Landi b, Bei Wang c,*
PMCID: PMC7755142  NIHMSID: NIHMS1650240  PMID: 33367228

Abstract

Topological data analysis and its main method, persistent homology, provide a toolkit for computing topological information of high-dimensional and noisy data sets. Kernels for one-parameter persistent homology have been established to connect persistent homology with machine learning techniques with applicability on shape analysis, recognition and classification. We contribute a kernel construction for multi-parameter persistence by integrating a one-parameter kernel weighted along straight lines. We prove that our kernel is stable and efficiently computable, which establishes a theoretical connection between topological data analysis and machine learning for multivariate data analysis.

Keywords: Topological Data Analysis, Machine Learning, Persistent Homology, Multivariate Analysis

1. Introduction

Topological data analysis (TDA) is an active area in data science with a growing interest and notable successes in a number of applications in science and engineering [1, 2, 3,4, 5, 6, 7, 8]. TDA extracts in-depth geometric information in amorphous solids [5], determines robust topological properties of evolution from genomic data sets [2] and identifies distinct diabetes subgroups [6] and a new subtype of breast cancer [9] in high-dimensional clinical data sets, to name a few. In the context of shape analysis, TDA techniques have been used in the recognition, classification [10, 11], summarization [12], and clustering [13] of 2D/3D shapes and surfaces. Oftentimes, such techniques capture and highlight structures in data that conventional techniques fail to treat [11, 13] or reveal properly [5].

TDA employs the mathematical notion of simplicial complexes [14] to encode higher order interactions in the system, and at its core uses the computational framework of persistent homology [15, 16, 17, 18, 19] to extract multi-scale topological features of the data. In particular, TDA extracts a rich set of topological features from high-dimensional and noisy data sets that complement geometric and statistical features, which offers a different perspective for machine learning. The question is, how can we establish and enrich the theoretical connections between TDA and machine learning?

Informally, homology was developed to classify topological spaces by examining their topological features such as connected components, tunnels, voids and holes of higher dimensions; persistent homology studies homology of a data set at multiple scales. Such information is summarized by the persistence diagram, a finite multi-set of points in the plane. A persistence diagram yields a complete description of the topological properties of a data set, making it an attractive tool to define features of data that take topology into consideration. Furthermore, a celebrated theorem of persistent homology is the stability of persistence diagrams [20] – small changes in the data lead to small changes of the corresponding diagrams, making it suitable for robust data analysis.

However, interfacing persistence diagrams directly with machine learning poses technical difficulties, because persistence diagrams contain point sets in the plane that do not have the structure of an inner product, which allows length and angle to be measured. In other words, such diagrams lack a Hilbert space structure for kernel-based learning methods such as kernel SVMs or PCAs [21]. Recent work proposes several variants of feature maps [22, 21, 23] that transform persistence diagrams into L2-functions over R2. This idea immediately enables the application of topological features for kernel-based machine learning methods as establishing a kernel function implicitly defines a Hilbert space structure [21].

A serious limit of standard persistent homology and its initial interfacing with machine learning [22, 21, 23, 24, 25] is the restriction to only a single scale parameter, thereby confining its applicability to the univariate setting. However, in many real-world applications, such as data acquisition and geometric modeling, we often encounter richer information described by multivariate data sets [26, 27, 28]. Consider, for example, climate simulations where multiple physical parameters such as temperature and pressure are computed simultaneously; and we are interested in understanding the interplay between these parameters. Consider another example in multivariate shape analysis, various families of functions carry information about the geometry of 3D shape objects, such as mesh density, eccentricity [29] or Heat Kernel Signature [30]; and we are interested in creating multivariate signatures of shapes from such functions. Unlike the univariate setting, very few topological tools exist for the study of multivariate data [31, 32, 29], let alone the integration of multivariate topological features with machine learning.

The active area of multi-parameter persistent homology [26] studies the extension of persistence to two or more (independent) scale parameters. A complete discrete invariant such as the persistence diagram does not exist for more than one parameter [26]. To gain partial information, it is common to study slices, that is, one-dimensional affine subspaces where all parameters are connected by a linear equation. In this paper, we establish, for the first time, a theoretical connection between topological features and machine learning algorithms via the kernel approach for multi-parameter persistent homology. Such a theoretical underpinning is necessary for applications in multivariate data analysis.

Our contribution.

We propose the first kernel construction for multi-parameter persistent homology. Our kernel is generic, stable and can be approximated in polynomial time. For simplicity, we formulate all our results for the case of two parameters, although they extend to more than two parameters.

Our input is a data set that is filtered according to two scale parameters and has a finite description size; we call this a bi-filtration and postpone its formal definition to Section 2. Our main contribution is the definition of a feature map that assigns to a bi-filtration X a function ΦX:Δ(2)R, where Δ(2) is a subset of R4. Moreover, ΦX2 is integrable over Δ(2), effectively including the space of bi-filtrations into the Hilbert space L2(2)). Therefore, based on the standard scalar product in L2(2)) a 2-parameter kernel is defined such that for two given bi-filtrations X and Y we have

X,YΦΔ(2)ΦXΦYdμ. (1)

We construct our feature map by interpreting a point of Δ(2) as a pair of (distinct) points in R2 that define a unique slice. Along this slice, the data simplifies to a mono-filtration (i.e., a filtration that depends on a single scale parameter), and we can choose among a large class of feature maps and kernel constructions of standard, one-parameter persistence. To make the feature map well-defined, we restrict our attention to a finite rectangle R.

Our inclusion into a Hilbert space induces a distance between bi-filtrations as

dΦ(X,Y)(ΦXΦY)2dμ. (2)

We prove a stability bound, relating this distance measure to the matching distance and the interleaving distance (see the paragraph on related work below). We also show that this stability bound is tight up to constant factors (see Section 4).

Finally, we prove that our kernel construction admits an efficient approximation scheme. Fixing an absolute error bound ϵ, we give a polynomial time algorithm in 1/ϵ and the size of the bi-filtrations X and Y to compute a value r such that rX,YΦr+ϵ. On a high level, the algorithm subdivides the domain into boxes of smaller and smaller width and evaluates the integral of (1) by lower and upper sums within each subdomain, terminating the process when the desired accuracy has been achieved. The technical difficulty lies in the accurate and certifiable approximation of the variation of the feature map when moving the argument within a subdomain.

Related work.

Our approach heavily relies on the construction of stable and efficiently computable feature maps for mono-filtrations. This line of research was started by Reininghaus et al. [21], whose approach we discuss in some detail in Section 2. Alternative kernel constructions appeared in [24, 33]. Kernel constructions fit into the general framework of including the space of persistence diagrams in a larger space with more favorable properties. Other examples of this idea are persistent landscapes [22] and persistent images [34], which can be interpreted as kernel constructions as well. Kernels and related variants defined on mono-filtrations have been used to discriminate and classify shapes and surfaces [21, 25]. An alternative approach comes from the definition of suitable (polynomial) functions on persistence diagrams to arrive at a fixed-dimensional vector in Rd on which machine learning tasks can be performed; see [35, 36, 37, 38].

As previously mentioned, a persistence diagram for multiparameter persistence does not exist [26]. However, bi-filtrations still admit meaningful distance measures, which lead to the notion of closeness of two bi-filtrations. The most prominent such distance is the interleaving distance [39], which, however, has recently been proved to be NP-complete to compute and approximate [40]. Computationally attractive alternatives are (multi-parameter) bottleneck distance [41] and the matching distance [42, 43], which compares the persistence diagrams along all slices (appropriately weighted) and picks the worst discrepancy as the distance of the bi-filtrations. This distance can be approximated up to a precision ϵ using an appropriate subsample of the lines [42], and also computed exactly in polynomial time [43]. Our approach extends these works in the sense that not just a distance, but an inner product on bi-filtrations, is defined with our inclusion into a Hilbert space. In a similar spirit, the software library RIVET [44] provides a visualization tool to explore bi-filtrations by scanning through the slices.

2. Preliminaries

We introduce the basic topological terminology needed in this work. We restrict ourselves to the case of simplicial complexes as input structures for a clearer geometric intuition of the concepts, but our results generalize to more abstract input types (such as minimal representations of persistence modules) without problems.

Mono-filtrations.

Given a vertex set V, an (abstract) simplex is a non-empty subset of V, and an (abstract) simplicial complex is a collection of such subsets that is closed under the operation of taking non-empty subsets. A subcomplex of a simplicial complex X is a simplicial complex Y with YX. Fixing a finite simplicial complex X, a mono-filtration X of X is a map that assigns to each real number α, a subcomplex X(α) of X, with the property that whenever αβ, X(α)X(β). The size of X is the number of simplices of X. Since X is finite, X(α) changes at only finitely many places when α grows continuously from −∞ to +∞; we call these values critical. More formally, α is critical if there exists no open neighborhood of α such that the mono-filtration assigns the identical subcomplex to each value in the neighborhood. For a simplex σ of X, we call the critical value of σ the infimum over all α for which σX(α). For simplicity, we assume that this infimum is a minimum, so every simplex has a unique critical value wherever it is included in the mono-filtration.

Bi-filtrations.

For points in R2, we write (a, b) ≤ (c, d) if ac and bd. Similarly, we say (a, b) < (c, d) if a < c and b < d. For a finite simplicial complex X, a bi-filtration X of X is a map that assigns to each point pR2 a subcomplex X(p) of X, such that whenever pq, X(p)X(q). Again, a point p = (p1, p2) is called critical for X if, for any ϵ > 0, both X(p1ϵ,p2) and X(p1,p2ϵ) are not identical to X(p). Note that unlike in the mono-filtration case, the set of critical points might not be finite. We call a bi-filtration tame if it has only finitely many such critical points. For a simplex σ, a point pR2 is critical for σ if, for any ϵ > 0, σ is neither in X(p1ϵ,p2) nor in X(p1,p2ϵ), whereas σ is in both X(p1+ϵ,p2) and X(p1,p2+ϵ). Again, for simplicity, we assume that σX(p) in this case. A consequence of tameness is that each simplex has a finite number of critical points. Therefore, we can represent a tame bi-filtration of a finite simplicial complex X by specifying the set of critical points for each simplex in X. The sum of the number of critical points over all simplices of X is called the size of the bi-filtration. We henceforth assume that bi-filtrations are always represented in this form; in particular, we assume tameness throughout this paper.

A standard example to generate bi-filtrations is by an arbitrary function F:XR2 with the property that if τσ are two simplices of X, F(τ) ≤ F(σ). We define the sublevel set XF(p) as

χF(p){σX|F(σ)p},

and let XF denote its corresponding sublevel set bi-filtration. It is easy to verify that XF yields a (tame) bi-filtration and F(σ) is the unique critical value of σ in the bi-filtration.

Slices of a bi-filtration.

A bi-filtration X contains an infinite collection of mono-filtrations. Let L be the set of all nonvertical lines in R2 with positive slope. Fixing any line L, we observe that when traversing this line in positive direction, the subcomplexes of the bi-filtration are nested in each other. Note that intersects the anti-diagonal x = −y in a unique base point b. Parameterizing as b + λ · a, where a is the (positive) unit direction vector of , we obtain the mono-filtration

X(α)X(b+αa).

We will refer to this mono-filtration X as a slice of X along (and sometimes also call itself the slice, abusing notation). The critical values of a slice can be inferred by the critical points of the bi-filtration in a computationally straightforward way. Instead of a formal description, we refer to Figure 1 for a graphical description. Also, if the bi-filtration is of size n, each of its slices is of size at most n.

Figure 1.

Figure 1.

The three black points mark the three critical points of some simplex σ in X. The shaded area denotes the positions at which σ is present in the bi-filtration. Along the given slice (red line), the dashed lines denote the first position where the corresponding critical point “affects” the slice. This position is either the upper-vertical, or right-horizontal projection of the critical point onto the slice, depending on whether the critical point is below or above the line. For σ, we see that it enters the slice at the position marked by the blue point.

Persistent homology.

A mono-filtration X gives rise to a persistence diagram. Formally, we obtain this diagram by applying the homology functor to X, yielding a sequence of vector spaces and linear maps between them, and splitting this sequence into indecomposable parts using representation theory. Instead of rolling out the entire theory (which is explained, for instance, in [45]), we give an intuitive description here.

Persistent homology measures how the topological features of a data set evolve when considered across a varying scale parameter α. The most common example involves a point cloud in Rd, where considering a fixed scale α means replacing the points by balls of radius α. As α increases, the data set undergoes various topological configurations, starting as a disconnected point cloud for α = 0 and ending up as a topological ball when α approaches ∞; see Figure 2(a) for an example in R2.

Figure 2.

Figure 2.

Computing persistent homology of a point cloud in R2. (a) A nested sequence of topological spaces formed by unions of balls at increasing parameter values. (b) A mono-filtration of simplicial complexes that captures the same topological information as in (a). (c) 0-dimensional and 1-dimensional persistence diagrams combined.

The topological information of this process can be summarized as a finite multi-set of points in the plane, called the persistence diagram. Each point of the diagram corresponds to a topological feature (i.e., connected components, tunnels, voids, etc.), and its coordinates specify at which scales the feature appears and disappears in the data. As illustrated in Figure 2(a), all five (connected) components are born (i.e., appear) at α = 0. The green component dies (i.e., disappears) when it merges with the red component at α = 2.5; similarly, the orange, blue and pink components die at scales 3, 3.2 and 3.7, respectively. The red component never dies as α goes to ∞. The 0-dimensional persistence diagram is defined to have one point per component with birth and death value as its coordinates (Figure 2(c)). The persistence of a feature is then merely its distance from the diagonal. While we focus on the components, the concept generalizes to higher dimensions, such as tunnels (1-dimensional homology) and voids (2-dimensional homology). For instance, in Figure 2(a), a tunnel appears at α = 4.2 and disappears at α = 5.6, which gives rise to a purple point (4.2, 5.6) in the 1-dimensional persistence diagram (Figure 2(c)).

From a computational point of view, the nested sequence of spaces formed by unions of balls (Figure 2(a)) can be replaced by a nested sequence of simplicial complexes by taking their nerves, thereby forming a mono-filtration of simplicial complexes that captures the same topological information but has a much smaller footprint (Figure 2(b)).

In the context of shape analysis, we apply persistent homology to capture the topological information of 2D and 3D shape objects by employing various types of mono-filtrations. A simple example is illustrated in Figure 3: we extract point clouds sampled from the boundary of 2D shape objects and compute the persistence diagrams using Vietoris-Rips complex filtrations.

Figure 3.

Figure 3.

The persistence diagrams of 2D shape objects. Black and red points are 0-dimensional and 1-dimensional features respectively (ignoring points with ∞ persistence).

Stability of persistent homology.

Bottleneck distance represents a similarity measure between persistence diagrams. Let D, D′ be two persistence diagrams. Without loss of generality, we can assume that both contain infinitely many copies of the points on the diagonal. The bottleneck distance between D and D′ is defined as

dB(D,D)infγsupxDxγ(x), (3)

where γ ranges over all bijections from D to D′. We will also use the notation dB(X,Y) for two mono-filtrations instead of dB(D(X),D(Y))

A crucial result for persistent homology is the stability theorem proven in [46] and re-stated in our notation as follows. Given two functions f,g:XR whose sublevel sets form two mono-filtrations of a finite simplicial complex X, the induced persistence diagrams satisfy

dB(Df,Dg)fgsupσX|f(σ)g(σ)|. (4)

Feature maps for mono-filtrations.

Several feature maps aimed at the construction of a kernel for mono-filtrations have been proposed in the literature [22, 23, 21]. We discuss one example: the persistence scale-space kernel [21] assigns to a mono-filtration X an L2-function ϕX defined on Δ(1){(x1,x2)R2x1<x2}. The main idea behind the definition of ϕX is to define a sum of Gaussian peaks, all of the same height and width, with each peak centered at one finite off-diagonal point of the persistence diagram D(X) of X. To make the construction robust against perturbations, the function has to be equal to 0 across the diagonal (the boundary of Δ(1)). This is achieved by adding negative Gaussian peaks at the reflections of the off-diagonal points along the diagonal. Writing z for the reflection of a point z, we obtain the formula,

ϕX(x)14πtzD(X)exz224texz¯224t, (5)

where t is the width of the Gaussian, which is a free parameter of the construction. See Figure 4 (b) and (c) for an illustration of a transformation of a persistence diagram to the function ϕX. The induced kernel enjoys several stability properties and can be evaluated efficiently without explicit construction of the feature map; see [21] for details.

Figure 4.

Figure 4.

An illustration of the construction of a feature map for multi-parameter persistent homology. (a) Given a bi-filtration X and a point (p, q) ∈ Δ(2), the line passing through them is depicted and the parameter λp and λq computed. (b) The point (λp,λq) is embedded in the persistence diagram of the mono-filtration X obtained as the slice of X along . (c) The point (λp, λq) is assigned the value ϕX(λp,λq) via the feature map ϕ.

More generally, in this paper, we look at the class of all feature maps that assign to a mono-filtration X a function in L2(1)). For such a feature map ϕX, we define the following properties:

  • Absolutely boundedness. There exists a constant v1 > 0 such that, for any mono-filtration X of size n and any x ∈ Δ(1), 0ϕX(x)v1n.

  • Lipschitzianity. There exists a constant v2 > 0 such that, for any mono-filtration X of size n and any x, x′ ∈ Δ(1), ϕX(x)ϕX(x)v2nxx2.

  • Internal stability. There exists a constant v3 > 0 such that, for any pair of mono-filtrations X, Y of size n and any x ∈ Δ(1), ϕX(x)ϕY(x)v3ndB(X,Y)

  • Efficiency. For any x ∈ Δ(1), ϕX(x) can be computed in polynomial time in the size of X, that is, in O(nk) for some k ≥ 0.

It can be verified easily that the scale-space feature map from above satisfies all these properties. The same is true, for instance, if the Gaussian peaks are replaced by linear peaks (that is, replacing the Gaussian kernel in (5) by a triangle kernel).

3. A feature map for multi-parameter persistent homology

Let ϕ be a feature map (such as the scale-space kernel) that assigns to a mono-filtration a function in L2(1)). Starting from ϕ, we construct a feature map Φ on the set of all bi-filtrations Ω that has values in a Hilbert space.

The feature map Φ assigns to a bi-filtration X a function ΦX:Δ(2)R. We set

Δ(2){(p,q)|p2,q2,p<q}

as the set of all pairs of points where the first point is smaller than the second one. Δ(2) can be interpreted naturally as a subset of R4, but we will usually consider elements of Δ(2) as pairs of points in R2.

Fixing (p, q) ∈ Δ(2), let denote the unique slice through these two points. Along this slice, the bi-filtration gives rise to a mono-filtration X, and consequently a function ϕX:Δ(1)R using the considered feature map for mono-filtrations. Moreover, using the parameterization of the slice as b + λ · a from Section 2, there exist real values λp, λq such that b + λpa = p and b + λqa = q. Since p < q and λp < λq, hence (λp, λq) ∈ Δ(1). We define ΦX(p,q) to be the weighted function value of ϕX at (λp, λq) (see also Figure 4), that is,

ΦX(p,q)w(p,q)ϕX(λp,λq), (6)

where w(p, q) is a weight function w:Δ(2)R defined below.

The weight function w has two components. First, let R be a bounded axis-aligned rectangle in R2; its bottom-left corner coincides with the origin of the coordinate axes. We define w such that its weight is 0 if p or q is outside of R. Second, for pairs of points within R × R, we assign a weight depending on the slope of the induced slices. Formally, let be parameterized as b + λ·a as above, and recall that a is a unit vector with non-negative coordinates. Write a = (a1, a2) and set ^min{a1,a2}. Then, we define

w(p,q)χR(p)χR(q)^,

where χR is the characteristic function of R, mapping a point x to 1 if xR and 0 otherwise.

The factor ^ ensures that slices that are close to being horizontal or vertical attain less importance in the feature map. The same weight is assigned to slices in the matching distance [42]. ^ is not important for obtaining an L2-function, but its meaning will become clear in the stability results of Section 4. We also remark that the largest weight is attained for the diagonal slice with a value of 12. Consequently, w is a non-negative function upper bounded by 12.

To summarize, our map Φ depends on the choice of an axis-aligned rectangle R and a choice of feature map for mono-filtrations, which itself might have associated parameters. For instance, using the scale-space feature map requires the choice of the width t (see (5)). It is only left to argue that the image of the feature map Φ is indeed an L2-function.

Theorem 1. If ϕ is absolutely bounded, then ΦX is in L2(2)).

Proof. Let X be a bi-filtration of size n. As mentioned earlier, each slice X is of a size at most n. By absolute boundedness and the fact that the weight function is upper bounded by 12, it follows that ΦX(p,q)v1n2 for all (p, q). Since the support of ΦX is compact (R × R), the integral of ΦX2 over Δ(2) is finite, being absolutely bounded and compactly supported. □

Note that Theorem 1 remains true even without restricting the weight function to R, provided we consider a weight function that is square-integrable over Δ(2). We skip the (easy) proof.

4. Stability

An important and desirable property for a kernel is its stability. In general, stability means that small perturbations in the input data imply small perturbations in the output data. In our setting, small changes between multi-filtrations (with respect to matching distance) should not induce large changes in their corresponding feature maps (with respect to L2 distance).

Adopted to our notation, the matching distance is defined as

dmatch(X,Y)=supL(^dB(X,Y)),

where L is the set of non-vertical lines with positive slope [47].

Theorem 2. Let X and Y be two bi-filtrations. If ϕ is absolutely bounded and internally stable, we have

ΦXΦYL2Cnarea(R)dmatch(X,Y),

for some constant C.

Proof. Absolute boundedness ensures that the left-hand side is well-defined by Theorem 1. Now we use the definition of ‖·‖L2 and the internal stability of ϕ to obtain

ΦXΦYL22=Δ(2)|w(p,q)ϕX(λp,λq)w(p,q)ϕY(λp,λq)|2dμΔ(2)(w(p,q)v3ndB(X,Y))2dμ=(v3n)2Δ(2)(w(p,q)dB(X,Y))2dμ

Since w(p, q) is zero outside R × R, the integral does not change when restricted to Δ(2) ∩ (R × R). Within this set, w(p, q) simplifies to ^, with the line through p and q. Hence, we can further bound

=(v3n)2Δ(2)(R×R)(^dB(X,Y))2dμ(v3n)2Δ(2)(R×R)supL(^dB(X,Y))2=dmatch(X,Y)dμ=(v3ndmatch(X,Y))2Δ(2)(R×R)1dμ.

The claimed inequality follows by noting that the final integral is equal to 14area(R)2. □

As a corollary, we get the the same stability statement with respect to interleaving distance instead of matching distance [48, Thm.1]. Furthermore, we obtain a stability bound for sublevel set bi-filtrations of functions XR2 [47, Thm.4]:

Corollary 3. Let F,G:XR2 be two functions that give rise to sublevel set bi-filtrations X and Y, respectively. If ϕ is absolutely bounded and internally stable, we have

ΦXΦYL2Cnarea(R)FG,

for some constant C.

We remark that the appearance of n in the stability bound is not desirable as the bound worsens when the complex size increases (unlike, for instance, the bottleneck stability bound in (4), which is independent of n). The factor of n comes from the internal stability property of ϕ, so we have to strengthen this condition on ϕ. However, we show that such an improvement is impossible for a large class of “reasonable” feature maps.

For two bi-filtrations X, Y we define XY by setting (XY)(p)X(p)Y(p) for all pR2. A feature map Φ is additive if ΦXY=Φ(X)+Φ(Y) for all bi-filtrations X, Y. Φ is called non-trivial if there is a bi-filtration X such that ‖Φ‖L2 ≠ 0. Additivity and non-triviality for feature maps ϕ on mono-filtrations is defined in the analogous way. Note that, for instance, the scale space feature map is additive. Moreover, because (XY)=XY for every slice , a feature map Φ is additive if the underlying ϕ is additive.

For mono-filtrations, no additive, non-trivial feature map ϕ can satisfy

ϕXϕYCnδdB(χ,y)

with X, Y mono-filtrations and δ ∈ [0, 1); the proof of this statement is implicit in [21, Thm 3]. With similar ideas, we show that the same result holds in the multi-parameter case.

Theorem 4. If Φ is additive and there exists C > 0 and δ ∈ [0, 1) such that

ΦXΦYL2Cnδdmatch(X,Y)

for all bi-filtrations X and Y, then Φ is trivial.

Proof. Assume to the contrary that there exists a bi-filtration X such that ΦXL2>0. Then, writing O for the empty bifiltration, by additivity we get Φi=1nXΦOL2=nΦXΦOL2>0. On the other hand, dmatch(i=1nX,O)=dmatch(X,O). Hence, with C and δ as in the statement of the theorem,

Φi=1nXΦOL2Cnδdmatch(i=1nX,O)=nΦXΦOL2Cnδdmatch(X,O)=n1δΦXΦOL2Cdmatch(X,O)n,

a contradiction. □

5. Approximability

We provide an approximation algorithm to compute the kernel of two bi-filtrations X and Y up to any absolute error ϵ > 0. Recall that our feature map Φ depends on the choice of a bounding box R. In this section, we assume R to be the unit square [0, 1] × [0, 1] for simplicity. We prove the following theorem that shows our kernel construction admits an efficient approximation scheme that is polynomial in 1/ϵ and the size of the bi-filtrations.

Theorem 5. Assume ϕ is absolutely bounded, Lipschitz, internally stable and efficiently computable. Given two bi-filtrations X and Y of size n and ϵ > 0, we can compute a number r such that rX,YΦr+ϵ in polynomial time in n and 1/ϵ.

The proof of Theorem 5 will be illustrated in the following paragraphs, postponing most of the technical details to Appendix A.

Algorithm. Given two bi-filtrations X and Y of size n and ϵ > 0, our goal is to efficiently approximate X,YΦ by some number r. On the highest level, we compute a sequence of approximation intervals (with decreasing lengths) J1, J2, J3,…, each containing the desired kernel value X,YΦ. The computation terminates as soon as we find some Ji of width at most ϵ, in which case we return the left endpoint as an approximation to r.

For sN (N being the set of natural numbers), we compute Js as follows. We split R into 2s × 2s congruent squares (each of side length 2s) which we refer to as boxes. See Figure 5(a) for an example when s = 3. We call a pair of such boxes a box pair. The integral from (1) can then be split into a sum of integrals over all 24s box pairs. That is,

X,YΦ=Δ(2)ΦXΦYdμ=(B1,B2)Δ(2)(B1×B2)ΦXΦYdμ.

For each box pair, we compute an approximation interval for the integral, and sum them up using interval arithmetic to obtain Js.

Figure 5.

Figure 5.

(a) The two given slices realize the largest and smallest possible slope among all slices traversing the pink box pair. It can be easily seen that the difference of the unit vector of the center line to one of the unit vectors of these two lines realizes A for the given box pair. (b) Computing variations for the center slice and a traversing slice of a box pair.

We first give some (almost trivial) bounds for X,YΦ. Let (B1, B2) be a box pair with centers located at c1 and c2, respectively. By construction, vol(B1 × B2) = 2−4s. By the absolute boundedness of ϕ, we have

Δ(2)(B1×B2)ΦXΦYdμ(B1×B2)(12v1n12v1n)dμ (7)
=v12n22vol(B1×B2)=v12n224s+1, (8)

where 12 is the maximal weight. Let Uv12n224s+1. If c1c2, then we can choose [0, U] as approximation interval. Otherwise, if c1c2, then Δ(2) ∩ (B1 × B2) = ∅; we simply choose [0, 0] as approximation interval.

We can derive a second lower and upper bound for X,YΦ as follows. We evaluate ΦX and ΦY at the pair of centers (c1, c2), which is possible due to the efficiency hypothesis of ϕ. Let vX=ΦX(c1,c2) and vY=ΦY(c1,c2). Then, we compute variations δX,δY0 relative to the box pair, with the property that, for any pair (p, q) ∈ B1 × B2, ΦX(p,q)[vXδX,vX+δX], and ΦY(p,q)[vYδY,vY+δY]. In other words, variations describe how far the value of ${ \Phi_{ \cal{ X \cal} _} $} (or ${ \Phi_{ \cal{ Y \cal} _} $}) deviates from its value at (c1, c2) within B1 × B2. Combined with the derivations starting in (7), we have for any pair (p, q) ∈ B1 × B2,

max{0,(vXδX)(vYδY)} (9)
ΦX(p,q)ΦY(p,q) (10)
min{v12n22,(vX+δX)(vY+δY)}. (11)

By multiplying the bounds obtained in (9) by the volume of Δ(2) ∩ (B1 × B2), we get a lower and an upper bound for the integral of ΦXΦY over a box pair (B1, B2). By summing over all possible box pairs, the obtained lower and upper bounds are the endpoints of Js.

Variations.

We are left with computing the variations relative to a box pair. For simplicity, we set δδX δ ≔ δx and explain the procedure only for X; the treatment of Y is similar.

We say that a slice traverses (B1, B2) if it intersects both boxes in at least one point. One such slice is the center slice c, which is the slice through c1 and c2. See Figure 5(b) for an illustration. We set D to be the maximal bottleneck distance of the center slice and every other slice traversing the box pair (to be more precise, of the persistence diagrams along the corresponding slices). We set W as the maximal difference between the weight of the center slice and any other slice traversing the box pair, where the weight w is defined as in Section 3. Write λc1 for the parameter value of c1 along the center slice. For every slice traversing the box pair and any point pB1, we have a value λp, yielding the parameter value of p along . We define L1 as the maximal difference of λp and λc1 among all choices of p and . We define L2 in the same way for B2 and set L ≔ max{L1, L2}. With these notations, we obtain Lemma 6 below.

Lemma 6. For all (p, q) ∈ B1 × B2,

|ΦX(p,q)ΦX(c1,c2)|v3n2D+v1nW+v2nL.

Proof. Plugging in (6) and using triangle inequality, we obtain

|ΦX(p,q)ΦX(c1,c2)|=|^ϕX(λp,λq)^cϕXc(λc1,λc2)|^|ϕX(λp,λq)ϕXc(λp,λq)|+ϕXc(λp,λq)|^^c|+^c|ϕXc(λp,λq)ϕXc(λc1,λc2)|

and bound the three parts separately. The first summand is upper bounded by v3nD2 because of internal stability of the feature map ϕ and because ^12 for any slice . The second summand is upper bounded by v1nW by the absolute boundedness of ϕ. The third summand is bounded by v2nL, because (λp,λq)(λc1,λc2)22(λp,λq)(λc1,λc2)L and by ϕ being Lipschitz, ϕXc(λp,λq)ϕXc(λc1,λc2)2v2nL and ^12. The result follows. ■

Next, we bound D by simple geometric quantities. We use the following lemma, whose proof appeared in [48]:

Lemma 7 ([48]). Let ℓ and ℓ′ be two slices with parameterizations b + λa and b′ + λa′, respectively. Then, the bottleneck distance of the two persistence diagrams along these slices is upper bounded by

2aa+bb^^.

We define A as the maximal infinity distance of the directional vector of the center slice c and any other slice traversing the box pair. We define B as the maximal infinity distance of the base point of c and any other . Finally, we set M as the minimal weight among all slices traversing the box pair. Using Lemma 7, we see that

D2A+BM^c,

and we set

δv3n(2A+B)2M^c+v1nW+v2nL. (12)

It follows from Lemma 6 and Lemma 7 that δ indeed satisfies the required variation property.

We remark that δ might well be equal to ∞, if the box pair admits a traversing slice that is horizontal or vertical, in which case the lower and upper bounds derived from the variation are vacuous. While (12) looks complicated, the values v1, v2, v3 are constants coming from the considered feature map ϕ, and all the remaining values can be computed in constant time using elementary geometric properties of a box pair. We only explain the computation of A in Figure 5(a) and skip the details of the other values.

Analysis.

At this point, we have not made any claim that the algorithm is guaranteed to terminate. However, its correctness follows at once because Js indeed contains the desired kernel value. Moreover, handling one box pair has a complexity that is polynomial in n, because the dominant step is to evaluate ΦX at the center (c1, c2). Hence, if the algorithm terminates at iteration s0, its complexity is

s=1s0O(24spoly(n)).

This is because in iteration s, 24s box pairs need to be considered. Clearly, the geometric series above is dominated by the last iteration, so the complexity of the method is O(24s0poly(n)). The last (and technically most challenging) step is to argue that s0=O(logn+log1ϵ), which implies that the algorithm indeed terminates and its complexity is polynomial in n and 1/ϵ.

To see that we can achieve any desired accuracy for the value of the kernel, i.e., that the interval width tends to 0, we observe that, if the two boxes B1, B2 are sufficiently far away and the resolution s is sufficiently large, the magnitudes A, B, W, and L in (12) are all small, because the parameterizations of two slices traversing the box pair are similar (see Lemmas 11, 12, 13 and 14 in Appendix A). Moreover, if every slice traversing the box pair has a sufficiently large weight (i.e., the slice is close to the diagonal), the value M in (12) is sufficiently large. These two properties combined imply that the variation of such a box pair (which we refer to as the good type) tends to 0 as s goes to ∞. Hence, the bound based on the variation tends to the correct value for good box pairs.

However, no matter how high the resolution, there are always bad box pairs for which either B1, B2 are close, or are far but close to horizontal and vertical, and hence yield a very large variation. For each of these box pairs, the bounds derived from the variation are vacuous, but we still have the trivial bounds [0, U] based on the absolute boundedness of ϕ. Moreover, the total volume of these bad box pairs goes to 0 when s goes to ∞ (see Lemma 9, Lemma 10 in Appendix A). So, the contribution of these box pairs tends to 0. These two properties complete the proof of Theorem 5.

A more careful investigation of our proof shows that the complexity of our algorithm is O(n80+k(1/ϵ)40), where k is the efficiency constant of the feature map (Section 2). We made little effort to optimize the exponents in this bound.

6. Conclusions and future developments

We restate our main results for the case of a multi-filtration X with d parameters: there is a feature map that associates to X a real-valued function ΦX whose domain is of dimension 2d, and introduces a kernel between a pair of multi-filtrations with a stable distance function, where the stability bounds depend on the (2d-dimensional) volume of a chosen bounding box. The proofs of these generalized results carry over from the results of this paper. Moreover, assuming that d is a constant, we claim that the kernel can be approximated in polynomial time to any constant (with the polynomial exponent depending on d). A proof of this statement requires to adapt the definitions and proofs of Appendix A to the higher-dimensional case; we omit details.

Other generalizations include replacing filtrations of simplicial complexes with persistence modules (with a suitable finiteness condition), passing to sublevel sets of a larger class of (tame) functions and replacing the scale-space feature map with a more general family of single-parameter feature maps. All these generalizations will be discussed in subsequent work.

The next step is an efficient implementation of our kernel approximation algorithm. We have implemented a prototype in C++, realizing a more adaptive version of the described algorithm. We have observed rather poor performance due to the sheer number of box pairs to be considered. Some improvements under consideration are to precompute all combinatorial persistence diagrams (cf. the barcode templates from [44]), to refine the search space adaptively using a quad-tree instead of doubling the resolution and to use techniques from numerical integration to handle real-world data sizes. We hope that an efficient implementation of our kernel will validate the assumption that including more than a single parameter will attach more information to the data set and improve the quality of machine learning algorithms using topological features.

Acknowledgments

This work was initiated at the Dagstuhl Seminar 17292 “Topology, Computation and Data Analysis”. We thank all members of the breakout session on multi-dimensional kernel for their valuable suggestions. RC, UF, and MK acknowledge support by the Austrian Science Fund (FWF) grant P 29984-N35. CL is partially supported by ARCES (University of Bologna), and FAR 2017 (University of Modena and Reggio Emilia). BW is partially supported by NSF grant DBI-1661375, IIS-1513616 and NIH grant R01-1R01EB022876-01. Fig. 3 is generated by Lin Yan.

Appendix A. Details on the Proof of Theorem 5

Overview.

Recall that our approximation algorithm produces an approximation interval Js for sN by splitting the unit square into 2s × 2s boxes. For notational convenience, we write u ≔ 2s for the side length of these boxes.

We would like to argue that the algorithm terminates after O(logn+log1ϵ) iterations, which means that after that many iterations, an interval of width ϵ has been produced. The following Lemma 8 gives an equivalent criterion in terms of u and n.

Lemma 8. Assume that there are constants e1, e2 > 0, such that width(Js) = O(ne1ue2). Then, width(Js0) ≤ ϵ for some s0=O(logn+log1ϵ).

Proof. Assume that width(Js) ≤ cne1ue2 for constants c and s sufficiently large. Since u = 2−s, a simple calculation shows that cne1 ue2ϵ if and only if slogc+e1logn+log1ϵe2. Hence, choosing

slogc+e1logn+log1ϵe2=O(logn+log1ϵ)

ensures that width(Js0) ≤ ϵ. □

In the rest of this section, we will show that width(Js) = O(n2 u0.1).

Classifying box pairs.

For the analysis, we partition the box pairs considered by the algorithm into 4 disjoint classes. We call a box pair (B1, B2):

  • null if c1c2,

  • close if c1c2 such that c1c22<u,

  • non-diagonal if c1c2 such that c1c22u and any line that traverses (B1, B2) satisfies ^<u15,

  • good if it is of neither of the previous three types.

According to this notation, the integral from (1) can then be split as

X,YΦ=X,Ynull+X,Yclose+X,Ynondiag+X,Ygood,

where, X,Ynull is defined Σ(B1,B2)nullΔ(2)(B1×B2)ΦXΦYdμ and analogously for the other ones. We let Js,null, Js,close, Js,non–diag, Js,good denote the four approximation intervals obtained from our algorithm when summing up the contributions of the corresponding box pairs. Then clearly, Js is the sum of these four intervals. For simplicity, we will write Jnull instead of Js,null when s is fixed, and likewise for the other three cases.

We observe first that the algorithm yields Jnull = [0, 0], so null box pairs can simply be ignored. Box pairs that are either close or non-diagonal are referred to as bad box pairs in Section 5. We proceed by showing that the width of Jclose, Jnon–diag, and Jgood are all bounded by O(n2u0.1).

Bad box pairs.

We start with bounding the width of Jclose. Let Bclose be the union of all close box pairs. Note that our algorithm assigns to each box pair (B1, B2) an interval that is a subset of [0, U]. Recall that U=v12n224s+1. U can be rewritten as v12n22vol(B1×B2), where vol(B1 × B2) is the 4-dimensional volume of the box pair (B1, B2). It follows that

width(Jclose)v12n22vol(Bclose). (A.1)

Lemma 9. For u12, vol(Bclose)4πu.

Proof. Fixed a point pR, for each point qR such that (p,q)Bclose and p < q, there exists a unique close box pair (B1, B2) that contains (p, q). By definition of close box pair, we have that:

pq2pc12+c1c22+c2q2u+2u.

Moreover, for u12, 2uu, and so pq22u. Equivalently, q belongs to the 2-ball B(p,2u) centered at p and of radius 2u. Then,

vol(Bclose)=Bclose1dμpRqB(p,2u)1dμpR4πudμ=4πu.

Consequently, combined with (A.1), we have

width(Jclose)4πv12n22u=O(n2u0.1).

Note that u < 1 and hence, uu0.1.

For the width of Jnon–diag, we use exactly the same reasoning, making use of the following Lemma 10. Let Bnon-diag be the union of all non-diagonal box pairs.

Lemma 10. For u252, vol(Bnon-diag)2u15.

Proof. Fixed a point pR, for each point qR such that (p,q)Bnon-diag and p < q, there exists a unique non-diagonal box pair (B1, B2) that contains (p, q). We have that q lies in:

  • Triangle T1(p) of vertices p = (p1, p2), (1, p2), and (1,p2+(1p1)a2a1), if the line of maximum slope passing through B1 × B2 is such that ^=a2 where a = (a1, a2) is the (positive) unit direction vector of ;

  • Triangle T2(p) of vertices p = (p1, p2), (p1, 1), and (p1+(1p2)a1a2,1), if the line of minimum slope passing through B1 × B2 is such that ^=a1 where a = (a1, a2) is the (positive) unit direction vector of .

Let us bound the area of the two triangles. Since the calculations are analogous, let us focus on T1(p). By definition, the basis of T1(p) is smaller than 1 while its height is bounded by a2a1. The maximum value for the height of T1(p) is achieved for a2=u15. So, by exploiting the identity a12+a22=1, we have

(a2a1)2=u251u25.

Under the conditions u252 and 12u251 we have

a2a12u15.

Therefore, area(T1(p))22u15. Similarly, area(T2(p))22u15. Finally,

vol(Bnon-diag)=Bnon-diag1dμpRqT1(p)T2(p)1dμpR2u15dμ2u15.

Good box pairs.

For good box pairs, we use the fact that the variation of a box pair yields a subinterval of [(vXδX)(vYδY)vol(B1×B2),(vX+δx)(vY+δY)vol(B1×B2)] as an approximation, so the width is bounded by 2(vXδY+vYδX)vol(B1×B2). Let Bgood be the union of all good box pairs. Since the volumes of all good box pairs sum up to at most one, that is, vol(Bgood)1, it follows that the width of Jgood is bounded by 2(vXδY+vYδX). By absolute boundedness, vX and vY are in O(n), and recall that by definition,

δX=v3n(2A+B)2M^c+v1nW+v2nL=O(n(A+BM2+W+L))

based on the fact that ^M. The same bound holds for δY. Hence,

width(Jgood)=O(n2(A+BM2+W+L)).

It remains to show that A+BM2+W+L=O(u0.1). Note that Mu15 because the box pair is assumed to be good. We will show in the next lemmas that A, B, W, and L are all in O(u), proving that the term is indeed in O(u0.1). This completes the proof of the complexity of the algorithm.

Lemma 11. Let (B1, B2) be a good box pair. Let a, a′ be the unit direction vectors of two lines that pass through the box pair. Then, aa2u. In particular, A=O(u).

Proof. Since (B1, B2) is a good box pair, the largest value for ‖aa′ is achieved when and ℓ′ correspond to the lines passing through the box pair(B1, B2) with minimum and maximum slope, respectively. By denoting as c1 = (c1,x, c1,y), c2 = (c2,x, c2,y) the centers of B1, B2, we define to be the line passing through the points c1+(u2,u2), c2+(u2,u2). Similarly, let us call ℓ′ the line passing through the points c1+(u2,u2), c2+(u2,u2). So, the unit direction vector a of can be expressed as

a=(c2+(u2,u2))(c1+(u2,u2))(c2+(u2,u2))(c1+(u2,u2))2.

Similarly, the unit direction vector a′ of ℓ′ is described by

a=(c2+(u2,u2))(c1+(u2,u2))(c2+(u2,u2))(c1+(u2,u2))2.

Then, by denoting as (·, ·) the scalar product,

aa2aa22=a22+a222a,a=2(1a,a)=2(1c2c1+(u,u)c2c1+(u,u)2,c2c1+(u,u)c2c1+(u,u)2)=2(1c2c1222u2c2c1+(u,u)2c2c1+(u,u)2).

By an elementary calculation, one can prove that

c2c1+(u,u)2c2c1+(u,u)2=4u2(u2+2(c2,xc1,x)(c2,yc1,y))+c2c124.

Then,

aa22(1c2c1222u24u2(u2+2(c2,xc1,x)(c2,yc1,y))+c2c124)=2(1+2u2c2c1224u2(u2+2(c2,xc1,x)(c2,yc1,y))+c2c124).

Since (B1, B2) is a good box pair, c2c12u. So,

aa22(1+2u14(u2+2(c2,xc1,x)(c2,yc1,y))+1).

Since 4(u2+2(c2,xc1,x)(c2,yc1,y))+11, we have that

aa22(1+2u1)=4u.

Therefore,

aa2u.

Lemma 12. Let (B1, B2) be a good box pair. Let = + b, ℓ′ = a′λ + b′ be two lines that pass through the box pair such that a, a′ are unit direction vectors and b, b′ are the intersection points with the diagonal of the second and the fourth quadrant. Then bb4u. In particular, B=O(u).

Proof. Since (B1, B2) is a good box pair, the largest value for ‖bb′ is achieved when and ℓ′ correspond to the lines passing through the box pair(B1, B2) with minimum and maximum slope, respectively. By denoting the centers of B1 and B2 by c1 and c2, we define to be the line passing through the points c1+(u2,u2), c2+(u2,u2). Similarly, let us call ℓ′ the line passing through the points c1+(u2,u2), c2+(u2,u2). So, can be expressed as

(x,y)=c2+(u2,u2)c1(u2,u2)c2+(u2,u2)c1(u2,u2)2t+c1+(u2,u2),

where t is a parameter running on R. By intersecting with the line y = −x, we get:

c2,x+u2c1,x+u2c2+(u2,u2)c1(u2,u2)2t+c1,xu2=c2,y+u2c1,y+u2c2+(u2,u2)c1(u2,u2)2tc1,yu2,

which can be written as

c1,x+c1,y=c1,x+c1,yc2,xc2,yc2+(u2,u2)c1(u2,u2)2t,

letting us deduce that

t=(c1,x+c1,y)c2+(u2,u2)c1(u2,u2)2c1,x+c1,yc2,xc2,y

So, by replacing t in the equation of we retrieve b:

b=c2+(u2,u2)c1(u2,u2)c2+(u2,u2)c1(u2,u2)2(c1,x+c1,y)c2+(u2,u2)c1(u2,u2)2c1,x+c1,yc2,xc2,y+c1+(u2,u2)=(u,u)(c1,x+c1,y)c1,x+c1,yc2,xc2,y+(c2c1)(c1,x+c1,y)c1,x+c1,yc2,xc2,y+c1+(u2,u2).

Similarly,

b=(u,u)(c1,x+c1,y)c1,x+c1,yc2,xc2,y+(c2c1)(c1,x+c1,y)c1,x+c1,yc2,xc2,y+c1+(u2,u2).

So,

bb=(2c1,x+c1,yc1,x+c1,yc2,xc2,y1)(u,u)=|c1,x+c1,y+c2,x+c2,yc2,x+c2,yc1,xc1,y|u,u4r|c2,x+c2,yc1,xc1,y|u.

Since (B1, B2) is a good box pair,

c2,x+c2,yc1,xc1,y=c2c11c2c12u.

Finally,

bb4uu=4u.

Lemma 13. Let (B1, B2) be a good box pair. Let ^, ^ be the weights of two lines ℓ and ℓ′ that pass through the box pair. Then ^^4u. In particular, W=O(u).

Proof. If ^=a1 and ^=a1, then, by applying Lemma 11,

|^^|=|a1a1|aa2u.

On the other hand, if ^=a1 and ^=a2, then there exists a line ℓ″ passing through the box pair (B1, B2) such that a=(22,22). By applying twice Lemma 11,

|^^|=|a1a2||a122|+|22a2|=|a1a1|+|a2a2|aa+aa4u.

The cases ^=a2, ^=a2 and ^=a2, ^=a1 can be treated analogously to the previous ones. □

Lemma 14. Let (p, q), (p′, q′) be two points in a good box pair (B1, B2) and let ℓ, ℓ′ be the lines passing through p, q and p′, q′, respectively. In accordance with the usual parametrization, we have that λpλp2u+4u and λqλq2u+4u. As a consequence, L=O(u).

Proof. Thanks to the definition of λp, the triangular inequality and Lemma 12, we have that:

λp=pb2pp2+pb2+bb22u+λp+4u.

So, we have that λpλp2u+4u, and, similarly, λpλp2u+4u. Then,

|λpλp|2u+4u.

Analogously, it can be proven that

|λqλq|2u+4u.

References

  • [1].Carstens CJ, Horadam KJ. Persistent homology of collaboration networks. Mathematical Problems in Engineering 2013;2013:7. [Google Scholar]
  • [2].Chan JM, Carlsson G, Rabadan R Topology of viral evolution. Proceedings of the National Academy of Sciences of the United States of America 2013;110(46):18566–18571. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [3].Giusti C, Pastalkova E, Curto C, Itskov V Clique topology reveals intrinsic geometric structure in neural correlations. Proceedings of the National Academy of Sciences of the United States of America 2015;112(44):13455–13460. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [4].Guo W, Banerjee AG. Toward automated prediction of manufacturing productivity based on feature selection using topological data analysis. In: IEEE International Symposium on Assembly and Manufacturing 2016, p. 31–36. [Google Scholar]
  • [5].Hiraoka Y, Nakamura T, Hirata A, Escolar EG, Matsue K, Nishiura Y Hierarchical structures of amorphous solids characterized by persistent homology. Proceedings of the National Academy of Sciences of the United States of America 2016;113(26):7035–7040. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [6].Li L, Cheng WY, Glicksberg BS, Gottesman O, Tamler R, Chen R, et al. Identification of type 2 diabetes subgroups through topological analysis of patient similarity. Science Translational Medicine 2015;7(311):311ra174. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [7].Wang Y, Ombao H, Chung MK. Topological data analysis of single-trial electroencephalographic signals. Annals of Applied Statistics 2017;12(3):1506–1534. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [8].Yoo J, Kim EY, Ahn YM, Ye JC. Topological persistence vineyard for dynamic functional brain connectivity during resting and gaming stages. Journal of Neuroscience Methods 2016;267(15):1–13. [DOI] [PubMed] [Google Scholar]
  • [9].Nicolau M, Levine AJ, Carlsson G Topology based data analysis identifies a subgroup of breast cancers with a unique mutational profile and excellent survival. Proceedings National Academy of Sciences of the United States of America 2011;108(17):7265–7270. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [10].Turner K, Mukherjee S, Boyer DM. Persistent homology transform for modeling shapes and surfaces. Information and Inference: A Journal of the IMA 2014;3(4):310–344. [Google Scholar]
  • [11].Li C, Ovsjanikov M, Chazal F Persistence-based structural recognition. In: IEEE Conference on Computer Vision and Pattern Recognition 2014, p. 2003–2010. [Google Scholar]
  • [12].Biasotti S, Falcidieno B, Spagnuolo M Extended Reeb graphs for surface understanding and description. In: Proceedings of the 9th International Conference on Discrete Geometry for Computer Imagery 2000, p. 185–197. [Google Scholar]
  • [13].Skraba P, Ovsjanikov M, Chazal F, Guibas L Persistence-based segmentation of deformable shapes. In: IEEE Computer Society Conference on Computer Vision and Pattern Recognition - Workshop on Non-Rigid Shape Analysis and Deformable Image Alignment 2010, p. 45–52. [Google Scholar]
  • [14].Munkres JR. Elements of algebraic topology. Redwood City, CA, USA: Addison-Wesley; 1984. [Google Scholar]
  • [15].Edelsbrunner H, Letscher D, Zomorodian AJ. Topological persistence and simplification. Discrete and Computational Geometry 2002;28:511–533. [Google Scholar]
  • [16].Zomorodian A, Carlsson G Computing persistent homology. Discrete and Computational Geometry 2005;33(2):249–274. [Google Scholar]
  • [17].Edelsbrunner H, Harer J Persistent homology - a survey. Contemporary Mathematics 2008;453:257–282. [Google Scholar]
  • [18].Edelsbrunner H, Harer J Computational Topology: An Introduction. Providence, RI, USA: American Mathematical Society; 2010. [Google Scholar]
  • [19].Edelsbrunner H, Morozov D Persistent homology: Theory and practice In: Proceedings of the European Congress of Mathematics. European Mathematical Society Publishing House; 2012, p. 31–50. [Google Scholar]
  • [20].Cohen-Steiner D, Edelsbrunner H, Harer J Stability of persistence diagrams. Discrete and Computational Geometry 2007;37(1):103–120. [Google Scholar]
  • [21].Reininghaus J, Huber S, Bauer U, Kwitt R A stable multi-scale kernel for topological machine learning. In: IEEE Conference on Computer Vision and Pattern Recognition 2015, p. 4741–4748. [Google Scholar]
  • [22].Bubenik P Statistical topological data analysis using persistence landscapes. The Journal of Machine Learning Research 2015;16(1):77–102. [Google Scholar]
  • [23].Kwitt R, Huber S, Niethammer M, Lin W, Bauer U Statistical topological data analysis - a kernel perspective In: Cortes C, Lawrence ND, Lee DD, Sugiyama M, Garnett R, editors. Advances in Neural Information Processing Systems 28. Curran Associates, Inc.; 2015, p. 3070–3078. [Google Scholar]
  • [24].Kusano G, Fukumizu K, Hiraoka Y Persistence weighted Gaussian kernel for topological data analysis. Proceedings of the 33rd International Conference on Machine Learning 2016;48:2004–2013. [Google Scholar]
  • [25].Hofer C, Kwitt R, Niethammer M, Uhl A Deep learning with topological signatures In: Guyon I, Luxburg UV, Bengio S, Wallach H, Fergus R, Vishwanathan S, et al. , editors. Advances in Neural Information Processing Systems 30. Curran Associates, Inc.; 2017, p. 1634–1644. [Google Scholar]
  • [26].Carlsson G, Zomorodian A The theory of multidimensional persistence. Discrete and Computational Geometry 2009;42(1):71–93. [Google Scholar]
  • [27].Carlsson G, Mémoli F Multiparameter hierarchical clustering methods In: Locarek-Junge H, Weihs C, editors. Classification as a Tool for Research. Springer Berlin; Heidelberg; 2010, p. 63–70. [Google Scholar]
  • [28].Chazal F, Cohen-Steiner D, Guibas LJ, Mémoli F, Oudot SY. Gromov-Hausdorff stable signatures for shapes using persistence. Eurographics Symposium on Geometry Processing 2009;28(5):1393–1403. [Google Scholar]
  • [29].Singh G, Mémoli F, Carlsson G Topological methods for the analysis of high dimensional data sets and 3D object recognition. In: Eurographics Symposium on Point-Based Graphics 2007, p. 91–100. [Google Scholar]
  • [30].Sun J, Ovsjanikov M, Guibas L A concise and provably informative multi-scale signature based on heat diffusion. Computer Graphics Forum 2009;28(5):1383–1392. [Google Scholar]
  • [31].Edelsbrunner H, Harer J Jacobi sets of multiple Morse functions In: Cucker F, DeVore R, Olver P, Süli E, editors. Foundations of Computational Mathematics, Minneapolis 2002. Cambridge University Press; 2002, p. 37–57. [Google Scholar]
  • [32].Edelsbrunner H, Harer J, Patel AK. Reeb spaces of piecewise linear mappings. In: Proceedings of the 24th Annual Symposium on Computational Geometry 2008, p. 242–250. [Google Scholar]
  • [33].Carriére M, Cuturi M, Oudot S Sliced Wasserstein kernel for persistence diagrams. In: Proceedings of the 34th International Conference on Machine Learning; vol. 70 2017, p. 664–673. [Google Scholar]
  • [34].Adams H, Chepushtanova S, Emerson T, Hanson E, Kirby M, Motta F, et al. Persistence images: A stable vector representation of persistent homology. The Journal of Machine Learning Research 2017;18(1):218–252. [Google Scholar]
  • [35].Adcock A, Rubin D, Carlsson G Classification of hepatic lesions using the matching metric. Computer Vision and Image Understanding 2014;121:36–42. [Google Scholar]
  • [36].Di Fabio B, Ferri M Comparing persistence diagrams through complex vectors. In: V., M E, P editors. International Conference on Image Analysis and Processing Springer, Cham; 2015, p. 294–305. [Google Scholar]
  • [37].Adcock A, Carlsson E, Carlsson G The ring of algebraic functions on persistence bar codes. Homology, Homotopy and Applications 2016;18:381–402. [Google Scholar]
  • [38].Kališnik S Tropical coordinates on the space of persistence barcodes. Foundations of Computational Mathematics 2018;19(1):1–29. [Google Scholar]
  • [39].Lesnick M The theory of the interleaving distance on multidimensional persistence modules. Foundations of Computational Mathematics 2015;15(3):613–650. [Google Scholar]
  • [40].Bjerkevik HB, Botnan MB, Kerber M Computing the interleaving distance is NP-hard. ArXiv: 1811.09165; 2018. [Google Scholar]
  • [41].Dey TK, Xin C Computing Bottleneck Distance for 2-D Interval Decomposable Modules. In: Proceedings of the 34th International Symposium on Computational Geometry 2018, p. 32:1–32:15. [Google Scholar]
  • [42].Biasotti S, Cerri A, Frosini P, Giorgi D A new algorithm for computing the 2-dimensional matching distance between size functions. Pattern Recognition Letters 2011;32(14):1735–1746. [Google Scholar]
  • [43].Kerber M, Lesnick M, Oudot S Exact computation of the matching distance on 2-parameter persistence modules. In: Proceedings of the 35th International Symposium on Computational Geometry 2019, p. 46:1–46:15. [Google Scholar]
  • [44].Lesnick M, Wright M Interactive visualization of 2-D persistence modules. arXiv:1512.00180; 2015. [Google Scholar]
  • [45].Oudot S Persistence theory: From Quiver Representation to Data Analysis; vol. 209 of Mathematical Surveys and Monographs. American Mathematical Society; 2015. [Google Scholar]
  • [46].Cohen-Steiner D, Edelsbrunner H, Harer J Stability of persistence diagrams. Discrete and Computational Geometry 2007;37(1):103–120. [Google Scholar]
  • [47].Biasotti S, Cerri A, Frosini P, Giorgi D, Landi C Multidimensional size functions for shape comparison. Journal of Mathematical Imaging and Vision 2008;32(2):161. [Google Scholar]
  • [48].Landi C The rank invariant stability via interleavings In: Chambers EW, Fasy BT, Ziegelmeier L, editors. Research in Computational Topology; vol. 13 Springer; 2018, p. 1–10. [Google Scholar]

RESOURCES