Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2023 May 1.
Published in final edited form as: Proc IEEE Int Conf Acoust Speech Signal Process. 2022 Apr 27;2022:5528–5532. doi: 10.1109/icassp43922.2022.9746382

SCATTERING STATISTICS OF GENERALIZED SPATIAL POISSON POINT PROCESSES

Michael Perlmutter , Jieqian He ‡,*, Matthew Hirn ‡,**
PMCID: PMC9460525  NIHMSID: NIHMS1829191  PMID: 36093040

Abstract

We present a machine learning model for the analysis of randomly generated discrete signals, modeled as the points of an inhomogeneous, compound Poisson point process. Like the wavelet scattering transform introduced by Mallat, our construction is naturally invariant to translations and reflections, but it decouples the roles of scale and frequency, replacing wavelets with Gabor-type measurements. We show that, with suitable nonlinearities, our measurements distinguish Poisson point processes from common self-similar processes, and separate different types of Poisson point processes.

Keywords: Scattering transform, Poisson point process, convolutional neural network

1. INTRODUCTION

Convolutional neural networks (CNNs) have obtained impressive results for a number of learning tasks in which the underlying signal data can be modelled as a stochastic process, including texture discrimination [1], texture synthesis [2, 3], time-series analysis [4], and wireless networks [5]. In many scenarios, it is natural to model the signal data as the points of a (potentially complex) spatial point process. Furthermore, there are numerous other fields, including stochastic geometry [6], forestry [7], geoscience [8] and genetics [9], in which spatial point processes are used to model the underlying generating process of certain phenomena (e.g., earthquakes). This motivates us to consider the capacity of CNNs to capture the statistical properties of such processes.

The Wavelet scattering transform [10] is a model for CNNs, which consists of an alternating cascade of linear wavelet transforms and complex modulus nonlinearities. It has provable stability and invariance properties and has been used to achieve near state of the art results in fields such as audio signal processing [11], computer vision [12], and quantum chemistry [13]. In this paper, we examine a generalized scattering transform that utilizes a broader class of filters (which includes wavelets). We primarily focus on filters with small support, which is similar to those used in most CNNs.

Expected wavelet scattering moments for stochastic processes with stationary increments were introduced in [14], where it is shown that such moments capture important statistical information of one-dimensional Poisson processes, fractional Brownian motion, α-stable Lévy processes, and a number of other stochastic processes. In this paper, we extend the notion of scattering moments to our generalized architecture, and generalize many of the results from [14]. However, the main contributions contained here consist of new results for more general spatial point processes, including inhomogeneous Poisson point processes, which are not stationary and do not have stationary increments. The collection of expected scattering moments is a non-parametric model for these processes, which we show captures important summary statistics.

In Section 2 we will define our expected scattering moments. Then, in Sections 3 and 4 we will analyze these moments for certain generalized Poisson point processes and self-similar processes. We will present numerical examples in Section 5, and provide a short conclusion in section 6.

2. EXPECTED SCATTERING MOMENTS

Let ψL2() be a compactly supported mother wavelet with dilations ψj(t) = 2jψ(2jt) for j, and let X(t),t, be a stochastic process with stationary increments. The first-order wavelet scattering moments are defined in [14] as SX(j)=Eψj*X, where the expectation does not depend on t since X(t) has stationary increments and Sγ,pX(t) = SX(γ, p) is a wavelet which implies X * ψj(t) is stationary. Much of the analysis of in [14] relies on the fact that these moments can be rewritten as SX(j)=Eψ¯j*dX, where dψ¯j=ψj. This motivates us to define scattering moments as the integration of a filter, against a random signed measure Y(dt).

To that end, let wL2d be a continuous window function with support contained in [0, 1]d. Denote by ws(t)=wts the dilation of w, and set gγ(t) to be the Gabor-type filter with scale s > 0 and central frequency ξd,

gγ(t)=ws(t)eiξt,γ=(s,ξ),td. (1)

Note that with an appropriately chosen window function w, (1) includes dyadic wavelet families in the case that s = 2j and |ξ| = C/s. However, it also includes many other filters, such as Gabor filters used in the windowed Fourier transform.

Let Y(dt) be a random signed measure and assume that Y is T-periodic for some T > 0 in the sense that for any Borel set B we have Y(B) = Y(B + Tei), for all 1 ⩽ id (where {ei}id is the standard orthonormal basis for d). For fL2d, set f*Y(t):=df(tu)Y(du). We define the first-order and second-order expected scattering moments, 1 ⩽ p, p′ > ∞, at location t as

Sγ,pY(t):=Egγ*Y(t)pand (2)
Sγ,p,γ,pY(t):=Egγ*Yp*gγ(t)p. (3)

Note Y(dt) is not assumed to be stationary, which is why these moments depend on t. Since Y(dt) is periodic, we may also define time-invariant scattering coefficients by

SY(γ,p):=1Td[0,T]dSγ,pY(t)dt,and
SYγ,p,γ,p:=1Td[0,T]dSγ,p,γ,pY(t)dt

In the following sections, we analyze these moments for arbitrary frequencies ξ and small scales s, thus allowing the filters gγ to serve as a model for the learned filters in CNNs. In particular, we will analyze the asymptotic behavior of the scattering moments as s decreases to zero.

3. SCATTERING MOMENTS OF GENERALIZED POISSON PROCESSES

In this section, we let Y(dt) be an inhomogeneous, compound spatial Poisson point process. Such processes generalize ordinary Poisson point processes by incorporating variable charges (heights) at the points of the process and a nonuniform intensity for the locations of the points. They thus provide a flexible family of point processes that can be used to model many different phenomena. In this section, we provide a review of such processes and analyze their first and second-order scattering moments.

Let λ(t) be a continuous, periodic function on d with

0<λmin:=inftλ(t)λ<, (4)

and define its first and second order moments by

mp(λ):=1Td[0,T]dλ(t)2dt,p=1,2.

A random measure N(dt):=j=1δtj(dt) is called an inhomogeneous Poisson point process with intensity function λ(t) if for any Borel set Bd,

P(N(B)=n)=eΛ(B)(Λ(B))nn!,Λ(B)=Bλ(t)dt,

and, in addition, N(B) is independent of N(B′) for all B′ that do not intersect B. Now let Ajj=1 be a sequence of i.i.d. random variables independent of N. An inhomogeneous, compound Poisson point process Y(dt) is given by

Y(dt)=j=1Ajδtj(dt). (5)

For a further overview of these processes, we refer the reader to Section 6.4 of [15].

3.1. First-order Scattering Asymptotics

Computing the convolution of gγ with Y(dt) gives

gγ*Y(t)=dgγ(tu)Y(du)=j=1Ajgγttj,

which can be interpreted as a waveform gγ emitting from each location tj. Invariant scattering moments aggregate the random interference patterns in |gγ * Y|. The results below show that the expectation of these interference patterns encode important statistical information related to the point process.

For notational convenience, we let

Λs(t):=Λ([ts,t]d)=[ts,t]dλ(u)du

denote the expected number of points of N in the support of gγ(t – ·). By conditioning on N ([ts, t]d), the number of points in the support of gγ, and using the fact that

[N([ts,t]d)>m]=Osdλm+1

one may obtain the following theorem.1

Theorem 1.

Let EA1p<, and λ(t) be a periodic continuous intensity function satisfying (4). Then for every td, every γ = (s, ξ) such that sdλ < 1, and every m ⩾ 1,

Sγ,pY(t)k=1meΛs(t)Λs(t)kk!Ej=1kAjwVjeisξVjp, (6)

where the error term ε(m, s, ξ, t) satisfies

|ε(m,s,ξ,t)|Cm,pλλminwppEA1pλm+1sd(m+1) (7)

and V1, V2, . . . is an i.i.d. sequence of random variables, independent of the Aj, taking values in the unit cube [0, 1]d and with density pV(v)=sdΛs(t)λ(tvs) for ν ∈ [0, 1]d.

If we set m = 1, and let s → 0, then one may use the fact that a small cube [ts, t]d has at most one point of N with overwhelming probability to obtain the following result.

Theorem 2.

Let Y(dt) satisfy the same assumptions as in Theorem 1. Let γk = (sk, ξk) be a sequence of scale and frequency pairs such that limk→∞ sk = 0. Then

limkSγk,pY(t)skd=λ(t)EA1pwpp, (8)

for all t, and consequently

limkSYγk,pskd=m1(λ)EA1pwpp. (9)

This theorem shows that for small scales the scattering moments Sγ,pY(t) encode the intensity function λ(t), up to factors depending upon the summary statistics of the charges Ajj=1 and the window w. Thus even a one-layer location-dependent scattering network yields considerable information regarding the underlying data generation process.

In the case of ordinary (non-compound) homogeneous Poisson processes, Theorem 2 recovers the constant intensity. For general λ(t) and invariant scattering moments, the role of higher-order moments of λ(t) is highlighted by considering higher-order expansions (e.g., m > 1) in (6). The next theorem considers second-order expansions and illustrates their dependence on the second moment of λ(t).

Theorem 3.

Let Y satisfy the same assumptions as in Theorem 1. If (γk)k⩾1 = (sk, ξk)k⩾1, is a sequence such that limk→∞ sk = 0 and limkskξk=Ld, then

limkSYγk,psk2dEA1pEVkp1Td[0,T]dΛsk(t)sk2ddt=m2(λ)EA1wU1eiLU1+A2wU2eiLU2p2wppEA1p, (10)

where U1, U2 are independent uniform random variables on [0, 1]d; and (Vk)k⩾1 is a sequence of random variables independent of the Aj taking values in the unit cube with respective densities, pVk(v)=skdΛsk(t)λtvsk for ν ∈ [0, 1]d.

We note that the scale normalization on the left hand side of (10) is s−2d, compared to a normalization of sd in Theorem 2. Thus, intuitively, (10) is capturing information at moderately small scales that are larger than the scales considered in Theorem 2. Unlike Theorem 2, which gives a way to compute m1(λ), Theorem 3 does not allow one to compute m2(λ) since it would require knowledge of Λsk(t) in addition to the distribution from which the charges Ajj=1 are drawn. However, Theorem 3 does show that at moderately small scales the invariant scattering coefficients depend nontrivially on the second moment of λ(t). Therefore, they can be used to distinguish between, for example, an inhomogeneous Poisson point process with intensity function λ(t) and a homogeneous Poisson point process with constant intensity.

3.2. Second-Order Scattering Moments of Generalized Poisson Processes

Our next result shows that second-order scattering moments encode higher-order moment information about the Ajj=1.

Theorem 4.

Let Y(dt) satisfy the same assumptions as in Theorem 1. Let γk = (sk, ξk) and γk=sk,ξk be sequences of scale-frequency pairs with sk=csk for some c > 0 and limkskξk=Ld. Let 1 ⩽ p, p′ < ∞ and q = pp′. Assume EA1q<, and let K:=gc,L/c*g1,0ppp. Then,

limkSγk,p,γk,pY(t)skdp+1=Kλ(t)EA1q,and (11)
limkSYγk,p,γk,pskdp+1=Km1(λ)EA1q. (12)

Theorem 2 shows first-order scattering moments with p = 1 are not able to distinguish between different types of Poisson point processes at very small scales if the charges have the same first moment. However, Theorem 4 shows second-order scattering moments encode higher-moment information about the charges, and thus are better able to distinguish them (when used in combination with the first-order coefficients). In Sec. 4, we will see first-order invariant scattering moments can distinguish Poisson point processes from self-similar processes if p = 1, but may fail to do so for larger values of p.

4. COMPARISON TO SELF-SIMILAR PROCESSES

We will show first-order invariant scattering moments can distinguish between Poisson point processes and certain self-similar processes, such as α-stable processes, 1 < α ⩽ 2, or fractional Brownian motion (fBM). These results generalize those in [14] both by considering more general filters and general pth scattering moments.

For a stochastic process X(t), t, we consider the convolution of the filter gγ with the noise dX defined by gγ*dX(t):=gγ(tu)dX(u), and define (in a slight abuse of notation) the first-order scattering moments at time t by Sγ,pX(t):=Egγ*dX(t)p. In the case where X(t) is a compound, inhomogeneous Poisson (counting) process, Y = dX will be a compound Poisson random measure and these scattering moments will coincide with those defined in (2).

The following theorem analyzes the small-scale first-order scattering moments when X is either an α-stable process, or an fBM. It shows the small-scale asymptotics of the corresponding scattering moments are guaranteed to differ from those of a Poisson point process when p = 1. We also note that both α-stable processes and fBM have stationary increments and thus Sγ,pX(t) = SX(γ, p) for all t.

Theorem 5.

Let 1p < ∞, and let γk = (sk, ξk) be a sequence of scale-frequency pairs with limk→∞ sk = 0 and limkskξk=L. Then, if X(t) is a symmetric α-stable process, p < α ⩽ 2, we have

limkSXγk,pskp/α=E01w(u)eiLudX(u)p.

Similarly, if X(t) is an fBM with Hurst parameter H ∈ (0, 1) and w has bounded variation on [0, 1], then

limkSXγk,pskpH=E01w(u)eiLudX(u)p.

This theorem shows that first-order invariant scattering moments distinguish inhomogeneous, compound Poisson processes from both α-stable processes and fractional Brownian motion except in the cases where p = α or p = 1/H. In particular, these measurements distinguish Brownian motion, from a Poisson point process except in the case where p = 2.

5. NUMERICAL ILLUSTRATIONS

We carry out several experiments to numerically validate the previously stated results. In all of our experiments, we hold the frequency ξ constant while letting s decrease to zero.

Compound Poisson point processes with the same intensities:

We generated three homogeneous compound Poisson point processes, all with intensity λ(t) ≡ λ0 = 0.01, where the charges A1,j, A2,j, and A3,j are chosen so that A1,j = 1 uniformly, A2,j~N0,π2, and A3,j are Rademacher random variables. The charges of the three signals have the same first moment EAi,j=1 and different second moment with E[A1,j2]=E[A3,j2]=1 and E[A2,j2]=π2. As predicted by Theorem 2, Figure 1 shows first-order scattering moments will not be able to distinguish between the three processes with p = 1, but will distinguish the process with Gaussian charges from the other two when p = 2.

Fig. 1.

Fig. 1.

Fig. 1.

Fig. 1.

First-order invariant scattering moments of homogeneous compound Poisson point processes with the same intensity λ0 and different Ai. Left: Realizations of the process with arrival rates given by Top: Ai = 1 Middle: Ai are normal random variables Bottom: Ai are Rademacher random variables. Middle: Plots of normalized first-order scattering SY(s,ξ,1)sw1 moments with p = 1.Right: Plots of normalized first-order scattering SY(s,ξ,2)sw22 moments with p = 2.

Inhomogeneous, non-compound Poisson point processes:

We also consider an inhomogeneous, non-compound Poisson point processes with intensity function λ(t)=0.011+0.5sin2πtN (where we estimate Sγ,pY(t), by averaging over 1000 realizations). Figure 2 plots the scattering moments for the inhomogeneous process at different times, and shows they align with the true intensity function.

Fig. 2.

Fig. 2.

Fig. 2.

First-order scattering moments for inhomogeneous Poisson point processes. Left: Sample realization with λ(t)=0.011+0.5sin2πtN. Right: Time-dependent scattering moments Sγ,pY(t)swPp at t1=N4, t2=N2,t3=3N4. Note that the scattering coefficients at times t1, t2, t3 converges to λ(t1) = 0.015, λ(t2) = 0.01, λ(t3) = 0.005.

Poisson point process and self similar process:

We consider a Brownian motion compared to a Poisson point process with intensity λ = 0.01 and charges (A)j=110. Figure 3 shows the convergence rate of the first-order scattering moments can distinguish these processes when p = 1 but not when p = 2.

Fig. 3.

Fig. 3.

Fig. 3.

Fig. 3.

First-order invariant scattering moments for standard Brownian motion and Poisson point process. Left: Sample realizations Top: Brownian motion. Bottom: Ordinary Poisson point process. Middle: Normalized scattering moments SYpoisson(x,ξ,p)λEA1pwpp and SXBM(s,ξ,p)λE|Z|pwpp for Poisson and BM with p = 1. Right: The same but with p = 2.

6. CONCLUSION

We have constructed Gabor-filter scattering transforms for random measures on d. Our work is closely related to [14] but considers more general classes of filters and point processes (although we note that [14] provides a more detailed analysis of self-similar processes). In future work, it would be interesting to explore the use of these measurements for tasks such as, e.g., synthesizing new signals.

A. PROOF OF THEOREM 1

To prove Theorem 1 we will need the following lemma.

Lemma 1.

Let Z be a Poisson random variable with parameter λ. Then for all α, m, 0 < λ < 1, we have

EZα1{Z>m}=k=m+1eλλkk!kαCm,αλm+1.

Proof. For 0 < λ < 1 and k, e−λλk ⩽ 1. Therefore,

EZα1{Z>m}=k=m+1eλλkk!kα=λm+1k=0eλλk(k+m+1)!(k+m+1)αλm+1k=0(k+m+1)α(k+m+1)!=Cα,mλm+1.

The proof of Theorem 1. Recalling the definitions of Y(dt) and Sγ,pY(t), and setting Ns(t) = N([ts, t]d), we see

Sγ,pY(t)=E[st,t]dwtuseiξ(tu)Y(du)p=Ej=1Ns(t)Ajwttjseiξttjp,

where t1,t2,tNs(t) are the points N(t) in [ts, t]d. Conditioned on the event that Ns(t) = k, the locations of the k points on [ts, t]d are distributed as i.i.d. random variables Z1, . . . , Zk taking values in [ts, t]d with density

pZ(z)=λ(z)Λs(t),z[ts,t]d.

Therefore, the random variables

Vi:=tZis

take values in the unit cube [0, 1]d and have density

pV(v)=sdΛs(t)λ(tvs),v[0,1]d.

Note that in the special case that N is homogeneous, i.e. λ(t) ≡ λ0 is constant, the Vi are uniform random variables on [0, 1]d.

Therefore, computing the conditional expectation, we have for k ⩾ 1

Ej=1Ns(t)Ajwttjseiξttjp:Ns(t)=k (13)
=Ej=1kAjwVjeisξVjp
λλminkpEA1pwpp, (14)

where (14) follows from (i) the independence of the random variables Aj and Vj; (ii) the fact that for any sequence of i.i.d. random variables Z1, Z2, . . .,

En=1kZnpkp1En=1kZnp=kpEZ1p;

and (iii) the fact that

EwVip=[0,1]d|w(v)|ppV(v)dvλλminwpp.

Therefore, since Ns(t)=k=eΛs(t)Λs(t)k/k!,

Ej=1Ns(t)Ajwttjseiξttjp=
=k=0eΛs(t)Λs(t)kk!Ej=1Ns(t)Ajwttjseiξttjp:Ns(t)=k
=k=1eΛs(t)Λs(t)kk!Ej=1kAjwVjeisξVjp
=k=1meΛs(t)Λs(t)kk!Ej=1kAjwVjeisξVjp+ε(m,s,ξ,t),

where

ε(m,s,t,ξ):=k=m+1eΛs(t)Λs(t)kk!Ej=1kAjwVjeisξVjp.

By (14) and Lemma 1, if s is small enough so that Λs(t) ⩽ sd∥λ∥ < 1, then:

ε(m,s,ξ,t)=k=m+1eΛs(t)Λs(t)kk!Ej=1kAjwVjeisξVjpλλminEA1pwppk=m+1eΛs(t)Λs(t)kk!kpCm,pλλminEA1pwppΛs(t)m+1Cm,pλλminEA1pwppλm+1sd(m+1).

B. PROOF OF THEOREM 2

Proof. Let (sk, ξk) be a sequence of scale and frequency pairs such that limk→∞ sk = 0. Applying Theorem 1 with m = 1, we obtain:

Sγk,pY(t)skd=eΛsk(t)Λsk(t)skdEA1wV1,keisξV1,kp+ε1,sk,ξk,tskd=eΛsk(t)Λsk(t)skdEA1pEwV1,kp+ε1,sk,ξk,tskd,

where we write V1,k = V1 to emphasize the fact that the density of V1,k is:

pVk(v)=skdΛsk(t)λtvsk.

Using the error bound (7), we see that:

limkε1,sk,ξk,tskd=0.

Furthermore, since 0Λsk(t)skdλ, we observe that:

limkeΛsk(t)=1,

and by the continuity of λ(t),

limkΛsk(t)skd=limk1skdskt,tdλ(u)du=λ(t). (15)

Finally, by the continuity of λ(t), we see that

pVk(v)λλminandlimkpVk(v)=1,v[0,1]d. (16)

Therefore, by the bounded convergence theorem,

limkEwV1p=limk[0,1]d|w(v)|ppVk(v)dv=[0,1]d|w(v)|plimkpVk(v)dv=wpp.

That completes the proof of (8).

To prove (9), we assume that λ(t) is periodic with period T along each coordinate and again use Theorem 1 with m = 1 to observe,

SYsk,ξk,pskd=EA1p1Td[0,T]deΛsk(t)Λsk(t)skd×[0,1]d|w(v)|ppVk(v)dvdt+1Td[0,1]dε1,sk,ξk,tskddt.

By (7), the second integral converges to zero as k → 8. Therefore,

limkSYsk,ξk,pskd=EA1pwpp1Td[0,T]dλ(t)dt,

by the continuity of λ(t) and the bounded convergence theorem.

C. PROOF OF THEOREM 3

Proof. We apply Theorem 1 with m = 2 and obtain:

Sγk,pY(t) (17)
=eΛsk(t)Λsk(t)EA1pEwV1,kp (18)
+eΛsk(t)Λsk(t)22EA1wV1,keiskξkV1,k+A2wV2,keiskξkV2,kp (19)
+ε2,sk,ξk,t,

where Vi,k, i = 1,2, are random variables taking values on the unit cube [0, 1]d with densities,

pVk(v)=skdΛsk(t)λtvsk.

Dividing both sides in (18) by sk2dwppEA1p and subtracting Λsk(t)sk2dEwV1,kpwpp yields:

Sγk,pY(t)sk2dwppEA1pΛsk(t)sk2dEwV1,kpwpp (20)
=eΛsk(t)Λsk(t)Λsk(t)sk2dEwV1,kpwpp (21)
+eΛsk(t)Λsk(t)2sk2dEA1wV1,keiskξkV1,k+A2wV2,keiskξkV2,kp2wppEA1p
+ε2,sk,ξk,tsk2dwppEA1p.

Using the error bound (7),

limkε2,sk,ξk,tsk2dwppEA1p=0, (22)

at a rate independent of t. Recalling (16) from the proof of Theorem 2, we use the fact that limkpVk1 and the bounded convergence theorem to conclude,

limkE|A1wV1,keiskξkV1,k+A2wV2,keiskξkV2,k|p (23)
=EA1wU1eiLU1+A2wU2eiLU2p, (24)

where Ui, i = 1, 2, are uniform random variables on the unit cube and L=limkskξk. Similarly,

limkEwV1,kpwpp=1. (25)

Lastly, recalling that sk → 0 as k → 8 and using (15) from the proof of Theorem 2, we see

limkeΛsk(t)Λsk(t)Λsk(t)sk2d (26)
=limkΛsk(t)skdlimkeΛsk(t)1skd=λ(t)limkeΛsk(t)1skd=λ(t)2. (27)

Now we integrate both sides of (21) over [0, T]d and divide by Td. Taking the limit as k → 8, on the left hand side we get:

limk1Td[0,T]dSγk,pY(t)sk2dwppEA1pΛsk(t)sk2dEwV1,kpwppdt=limkSYsk,ξk,psk2dwppEA1pEwV1,kpwppTd[0,T]dΛsk(t)sk2ddt=limkSYsk,ξk,psk2dEwV1,kpEA1p1Td[0,T]dΛsk(t)sk2ddt,

where we used the definition of the invariant scattering moments and (25). On the right hand side of (21), we use (25), (27) and the dominated convergence theorem to see that the first term is:

limk1Td[0,T]deΛsk(t)Λsk(t)Λsk(t)sk2dEwV1,kpwppdt=limk1Td[0,T]deΛsk(t)Λsk(t)Λsk(t)sk2ddt=1Td[0,T]dλ(t)2dt.

Using (15), (23), and the bounded convergence theorem, the second term of (21) is:

limk1Td[0,T]deΛsk(t)Xkdt=EA1wU1eiLU1+A2wU2eiLU2p2TdwppE|A1|p[0,T]dλ(t)2dt

where

Xk=EA1wV1,keiskξkV1,k+A2wV2,keiskξkV2,kp2wppEA1p.

Finally, the third term of (21) goes to zero using the bounded convergence theorem and (22). Putting together the left and right hand sides of (21) with these calculations finishes the proof.

D. PROOF OF THEOREM 4

Proof. As in the proof of Theorem 1, let Ns(t) = N ([ts, t]d) denote the number of points in the cube [ts, t]d. Then since the support of w is contained in [0, 1]d,

gγk*Y(t)=tsk,tdwtuskeiξk(tu)Y(du)=j=1Nsk(t)Ajwttjskeiξkttj,

where t1,t2,,tNsk(t) are the points of N in [tsk, t]d. Therefore, in the event that Nsk(t)=1,

gγk*Y(t)p=gγkp*|Y|p(t),

and so, partitioning the space of possible outcomes based on Nsk(t), we obtain:

gγk*Y(t)p=gγk*Y(t)1Nsk(t)=1+gγk*Y(t)1Nsk(t)>1p=gγk*Y(t)1Nsk(t)=1p+gγk*Y(t)1Nsk(t)>1p=gγkp*|Y|p(t)1Nsk(t)=1+gγk*Y(t)1Nsk(t)>1p=gγkp*|Y|p(t)+ek(t),

where

ek(t):=gγk*Y(t)1Nsk(t)>1pgγkp*|Y|p(t)1Nsk(t)>1

Using the above, we can write the second order convolution term as:

gγk*gγk*Y(t)=gγk*gγkp*|Y|p(t)+gγk*ek(t).

The following lemma implies that gγk*ek(t) decays rapidly in Lp at a rate independent of t.

Lemma 2.

There exists δ > 0, independent of t, such that if sk < δ,

Egγk*ek(t)pCp,p,w,c,Lλλminλ2skdp+2.

Once we have proved Lemma 2, equation (11) will follow once we show,

limkEgγk*gγkp*|Y|p(t)pskdp+1=Kp,p,w,c,Lλ(t)EA1q. (28)

Let us prove (28) first and postpone the proof of Lemma 2. We will use the fact that the support of gγk*gγkp is contained in 0,sk+skd. Let s˜k:=sk+sk, Nk(t):=Ns¯k(t), Λk(t):=Λs¯k(t), and let t1,t2,,tNk(t) be the points of N in the cube ts˜k,td. We have that Nk(t)=n=eΛk(t)Λk(t)nn!, and conditioned on the event that Nk(t) = n, the locations of the points t1, . . . , tn are distributed as i.i.d. random variables Z1(t), . . . , Zn(t) taking values in ts˜k,td with density pZ(t)(z)=λ(z)Λk(t). Therefore the i.i.d. random variables V˜1(t),,V˜n(t) defined by V˜i(t):=tZi(t) take values in 0,s˜kd and have density

pV˜(t)(v)=λ(tv)Λk(t),v0,s˜kd.

Now, we condition on Nk(t) to see that

Egγk*gγkp*|Y|p(t)p (29)
=Ej=1Nk(t)Ajpgγk*gγkpttjp
=n=1eΛk(t)Λk(t)nn! (30)
Ej=1Nk(t)Ajpgγk*gγkpttjp:Nk(t)=n
=n=1eΛk(t)Λk(t)nn!Ej=1nAjpgγk*gγkpV˜j(t)p
=eΛk(t)Λk(t)EA1qEgγk*gγkpV˜1(t)p (31)
+n=2eΛk(t)Λk(t)nn!Ej=1nAjpgγk*gγkpV˜j(t)p (32)

The following lemma will be used to estimate the scaling of the term in (31).

Lemma 3.

For all td,

limks˜kdskdp+1Egγk*gγkpV˜1(t)p=gc,L/c*g1,0ppp. (33)

Furthermore, there exists δ > 0, independent of t, such that if sk < δ then

s˜kdskdp+1Egγk*gγkpV˜1(t)p2λλminCp,p,w,c,L. (34)

Proof. Making a change of variables in both u and v, and recalling the assumption that sk=csk, we observe that

s˜kdskdp+1Egγk*gγkpV˜1(t)p=s˜kdskdp+1dpV˜(t)(v). (35)
dwvuskeiξk(vu)wuskpdupdv=s˜kddpV˜(t)skvdwsk(vu)skeiskξk(vu)w(u)pdupdv=ds˜kdλtskvΛk(t)dwuvceiskξk(uv)/cw(u)pdupdv. (36)

The continuity of λ(t) implies that

limks˜kdλtskvΛk(t)=1,v[0,1+c]d.

Furthermore, the assumption 0 < λmin ⩽ ∥λ < ∞ implies

s˜kdλtskvΛk(t)λλmin,k1. (37)

Therefore, (33) follows from the dominated convergence theorem and by the observation that the inner integral of (36) is zero unless v ∈ [0, 1 + c]d. Equation (34) follows from inserting (37) into (36) and sending k to infinity.

Since

limkΛk(t)s˜kd=λ(t),

the independence of V˜1(t) and A1, the continuity of λ(t), and Lemma 3 imply that taking k → ∞ in (31) yields:

limkeΛk(t)Λk(t)EA1qEgγk*gγkpV˜1(t)pskdp+1=limkeΛk(t)Λk(t)s˜kdEA1qs˜kdskdp+1Egγk*gγkpV˜1(t)p=Kp,p,c,w,Lλ(t)EA1q.

The following lemma shows that (32) is Oskdp+2 (and converges at a rate independent of t), and therefore completes the proof of (11) subject to proving Lemma 2.

Lemma 4.

For all α there exists δ > 0, independent of t, such that if sk < δ, then

n=2eΛk(t)Λk(t)nn!nαEj=1nAjpgγk*gγkpV˜j(t)pCp,p,w,c,α,Lλλminλ2EA1qskdp+2.

Proof. For any sequence of i.i.d. random variables, Z1, Z2, . . . , it holds that

En=1kZnpkp1En=1kZnp=kpEZ1p.

Therefore, by Lemma 1, Lemma 3, and the fact that the V˜j(t) and Ai are i.i.d. and independent of each other, we see that if sk < δ, where δ is as in (34),

n=2eΛk(t)Λk(t)nn!nα
×Ej=1nAjpgγk*gγkpV˜j(t)p
n=2eΛk(t)Λk(t)nn!nαnp
×EA1qgγk*gγkpV˜1(t)p
=n=2eΛk(t)Λk(t)nn!np+α
×EA1qEgγk*gγkpV˜1(t)p
=EA1qEgγk*gγkpV˜1(t)p
×n=2eΛk(t)Λk(t)nn!np+α
Cp,p,w,c,LλλminEA1qskdp+1s˜kd
×n=2eΛk(t)Λk(t)nn!np+α
Cp,p,w,c,L,αλλminEA1qskdp+1s˜kdΛk(t)2
Cp,p,w,c,L,αλλminλ2EA1qskdp+2,

where the last inequality uses the fact that Λk(t)s˜kdλ=(1+c)dskdλ.

We will now complete the proof of the theorem by proving Lemma 2.

Proof. [Lemma 2] Since

ek(t)=gγk*Y(t)1Nsk(t)>1pgγkp*|Y|p(t)1Nsk(t)>1,

we see that

gγk*ek(t)gγk*gγk*Y1Nsk()>1p(t)+gγk*gγkp*|Y|p1Nsk()>1(t).

First turning our attention to the second term, we note that

gγk*gγkp*|Y|p1Nsk()>1(t)=tsk,tdwtuskeiξk(tu)gγkp*|Y|p(u)1Nsk(u)>1du1Nk(t)>1tsk,tdwtuskgγkp*|Y|p(u)du=1Nk(t)>1gsk,0*gγkp*|Y|p(t). (38)

since Nsk(u)Nsk+sk(t)=Ns¯k(t)=Nk(t) for all utsk,td. Therefore, conditioning on Nk(t), if sk < δ,

Egγk*gγkp*|Y|p1Nsk()>1(t)p
E1Nk(t)>1gsk,0*gγkp*|Y|p(t)p
=n=2eΛk(t)Λk(t)nn!Ej=1nAjpgsk,0*gγkpV˜j(t)p
Cp,p,w,c,Lλλminλ2EA1qskdp+2

by Lemma 4. Now, turning our attention to the first term, note that

gγk*Y(t)p1Nsk(t)>1Nsk(t)p1gγkp*|Y|p(t)1Nsk(t)>1.

Therefore, by the same logic as in (38)

gγk*gγk*Y1Nsk()>1p(t)tsk,tdwtuskNsk(u)p1gγkp*|Y|p(u)1Nsk(u)>1du1Nk(t)>1Nk(t)p1tsk,tdwtuskgγkp*|Y|p(u)du1Nk(t)>1Nk(t)p1gsk,0*gγkp*|Y|p(t).

So again conditioning on Nk (t), and applying Lemma 4, we see that if sk < δ

Egγk*gγk*Y1Nsk()>1p(t)pn=2eΛk(t)Λk(t)kn!np1Ej=1nAjpgsk,0*gγkV˜j(t)ppCp,p,w,c,Lλλminλ2EA1qskdp+2.

This completes the proof of (11). Line (12) follows from integrating with respect to t, observing that the error bounds in Lemmas 2 and 3 are independent of t, and applying the bounded convergence theorem.

E. THE PROOF OF THEOREM 5

In order to prove Theorems 5, we will need the following lemma which shows that the scaling relationship of a self-similar process X(t) induces a similar relationship on stochastic integrals against dX(t).

Lemma 5.

Let X be a stochastic process that satisfies the scaling relation

X(st)=dsβX(t) (39)

for some β > 0 (where =d denotes equality in distribution). Then for any measurable function f:,

0sf(u)dX(u):=sβ01f(su)dX(u).

Proof. Let X=(X(t))t be a stochastic process satisfying (39), and let Pn={0=t0n<t1n<<tKnn=1} be a sequence of partitions of r0,1s such that

limnmaxktkntk1n=0.

Then, by the scaling relation (39),

0sf(u)dX(u)=limnk=0Kn1fstknXstk+1nXstkn:=sβlimnk=0Kn1fstknXtk+1nXtkn=sβ01f(su)dX(u).

We will now use Lemma 5 to prove Theorem 5.

Proof. We first consider the case where X=(X(t))t is an α-stable process, p < α ⩽ 2. Since X has stationary increments, its scattering coefficients do not depend on t and it suffices to analyze

Egγk*dX(0)p=Esk0gγk(u)dX(u)p=E0skgγk(u)dX(u)p,

where the second equality uses the fact the distribution of X does not change if it is run in reverse, i.e.

(X(t))t=d(X(t))t

It is well known that X(t) satisfies (39) for β = 1/α. Therefore, by Lemma 5

Egγk*dX(0)p=E0skwuskeiξkudX(u)p=skp/αE01w(u)eiξkskudX(u)p.

So,

Egγk*dX(0)pskp/α=E01w(u)eiξkskudX(u)p.

The proof will be complete as soon as we show that

limkE01w(u)eiξkskudX(u)p1/p=E01w(u)eiLudX(u)p1/p.

By the triangle inequality,

E01w(u)eiξkskudX(u)p1/pE01w(u)eiLudX(u)p1/pE01w(u)eiξkskueiLudX(u)p1/p.

Since 1 ⩽ p < α, we may choose p′ strictly greater than 1 such that pp′ < α, and note that by Jensen’s inequality

E01w(u)eiξkskueiLudX(u)p1/pE01w(u)eiξkskueiLudX(u)p1/p,

and since X(t) is a p′-integrable martingale, the boundedness of martingale transforms (see [16] and also [17]) implies

E01w(u)eiξkskueiLudX(u)p1/pCpsup0u1w(u)eiξkskueiLuEX1pCpskξkLwEX1p,

which converges to zero by the continuity of w on [0, 1] and the assumption that skξk converges to L.

Similarly, in the case where (X(t))t is a fractional Brownian motion with Hurst parameter H, we again need to show

limkE01w(u)eiξkskueiLudX(u)p1/p=0.

However, fractional Brownian motion is not a semi-martingale so we cannot apply Burkholder’s theorem as we did in the proof of Theorem 5. Instead, we use the Young-Lóeve estimate [18] which states that if xpuq is any (deterministic) function with bounded variation, and y(u) is any function which is α-Hölder continuous, 0 < α < 1, then

01x(u)dy(u)

is well-defined as the limit of Riemann sums and

01x(u)dy(u)x(0)(y(1)y(0))CαxBVyα,

where ∥·∥BV and ∥·∥α are the bounded variation and α-Hölder seminorms respectively. For all k, the function hk(u):=w(u)eiξkskueiLu:=w(u)fk(u) satisfies, hk(0) = 0 and

hkBVwfkBV+wBVfk.

One can check that the fact that skξk converges to L implies that fk converges to zero in both L and in the bounded variation seminorm, and that therefore that ∥hkBV converges to zero.

It is well-known that fractional Brownian motion with Hurst parameter H admits a continuous modification which is α-Hölder continuous for any α < H. Therefore,

E01w(u)eiξkskueiLudX(u)pCαphkBVpEXαp.

Lastly, one can use the Garsia-Rodemich-Rumsey inequality [19], to show that

EXαp<.

for all 1 < p < ∞. For details we refer the reader to the survey article [20]. Therefore,

limk0E01w(u)eiξkskueiLudX(u)p=0

as desired.

Remark 1.

The assumption that w has bounded-variation was used to justify that the stochastic integral against fractional Brownian motion was well defined as the limit of Riemann sums because of its Hölder continuity and the above mentioned result of [18]. This allowed us to avoid the technical complexities of defining such an integral using either the Malliavin calculus or the Wick product.

F. DETAILS OF NUMERICAL EXPERIMENTS

graphic file with name nihms-1829191-f0009.jpg

F.1. Definition of Filters

For all the numerical experiments, we take the window function w to be the smooth bump function

w(t)=exp14t4t2,t(0,1)0,otherwise.

Therefore for γ = (s, ξ), our filters are given by

gγ(t)=eiξtw(t)=eiξtes2/4ts4t2,t(0,s)0,otherwise.

F.2. Frequencies

In all of our experiments, we hold the frequency, ξ, which we sample uniformly at random from (0, 2π), constant while allowing the scale to decrease to zero.

F.3. Simulation of Poisson point process

We use the standard method to generate a realization of a Poisson point process. For Poisson point process with intensity λ, the time interval between two neighbor jumps follows exponential distribution:

Δj:=tjtj1~Exp(λ).

Therefore, taking the inverse cumulative distribution function, we sample the time interval between two neighbor jumps through:

Δj=logUjλ,

where Uj are i.i.d. uniform random variables on [0, 1], and assign the charge Aj to the jump at location tj.

For inhomogeneous Poisson process with intensity funciton λ(t), we simulate the time interval based on a well-known algorithm. We, first define the cumulated intensity:

Λ(t)=0tλ(s)ds,

then generate the location of jumps tj by the Algorithm 1.

Acknowledgments

M.H. acknowledges support from NSF DMS #1845856 (partially supporting J.H.), NIH NIGMS #R01GM135929, and DOE #DE-SC0021152.

Footnotes

1

A proof of Theorem 1, as well as the proofs of other theorems stated in this paper, can be found in the appendix

7. REFERENCES

  • [1].Sifre Laurent and Mallat Stéphane, “Rotation, scaling and deformation invariant scattering for texture discrimination,” in The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2013. [Google Scholar]
  • [2].Gatys Leon, Ecker Alexander S, and Matthias Bethge, “Texture synthesis using convolutional neural networks,” in Advances in Neural Information Processing Systems 28, 2015, pp. 262–270. [Google Scholar]
  • [3].Antognini Joseph, Hoffman Matt, and Weiss Ron J., “Synthesizing diverse, high-quality audio textures,” arXiv:1806.08002, 2018. [Google Scholar]
  • [4].Binkowski Mikolaj, Marti Gautier, and Donnat Philippe, “Autoregressive convolutional neural networks for asynchronous time series,” in Proceedings of the 35th International Conference on Machine Learning, Jennifer Dy and Andreas Krause, Eds., Stockholmsmässan, Stockholm Sweden, 10–15 Jul 2018, vol. 80 of Proceedings of Machine Learning Research, pp. 580–589, PMLR. [Google Scholar]
  • [5].Brochard Antoine, Bartłomiej Błaszczyszyn, Stéphane Mallat, and Zhang Sixin, “Statistical learning of geometric characteristics of wireless networks,” arXiv:1812.08265, 2018. [Google Scholar]
  • [6].Haenggi Martin, Andrews Jeffrey G., ois Baccelli Franc, Olivier Dousse, and Massimo Franceschetti, “Stochastic geometry and random graphs for the analysis and design of wireless networks,” IEEE Journal on Selected Areas in Communications, vol. 27, no. 7, pp. 1029–1046, 2009. [Google Scholar]
  • [7].Genet Astrid, Grabarnik Pavel, Sekretenko Olga, and Pothier David, “Incorporating the mechanisms underlying inter-tree competition into a random point process model to improve spatial tree pattern analysis in forestry,” Ecological Modelling, vol. 288, pp. 143–154, 09 2014. [Google Scholar]
  • [8].Schoenberg Frederic Paik, “A note on the consistent estimation of spatial-temporal point process parameters,” Statistica Sinica, 2016. [Google Scholar]
  • [9].Fromion V, Leoncini E, and Robert P, “Stochastic gene expression in cells: A point process approach,” SIAM Journal on Applied Mathematics, vol. 73, no. 1, pp. 195–211, 2013. [Google Scholar]
  • [10].Mallat Stéphane, “Group invariant scattering,” Communications on Pure and Applied Mathematics, vol. 65, no. 10, pp. 1331–1398, October 2012. [Google Scholar]
  • [11].Joakim Andén and Stéphane Mallat, “Deep scattering spectrum,” IEEE Transactions on Signal Processing, vol. 62, no. 16, pp. 4114–4128, August 2014. [Google Scholar]
  • [12].Oyallon Edouard and Mallat Stéphane, “Deep rototranslation scattering for object classification,” in Proceedings in IEEE CVPR 2015 conference, 2015, arXiv:1412.8659. [Google Scholar]
  • [13].Brumwell Xavier, Sinz Paul, Kim Kwang Jin, Qi Yue, and Hirn Matthew, “Steerable wavelet scattering for 3D atomic systems with application to Li-Si energy prediction,” in NeurIPS Workshop on Machine Learning for Molecules and Materials, 2018. [Google Scholar]
  • [14].Bruna Joan, Mallat Stéphane, Bacry Emmanuel, and Muzy Jean-Francois, “Intermittent process analysis with scattering moments,” Annals of Statistics, vol. 43, no. 1, pp. 323 – 351, 2015. [Google Scholar]
  • [15].Daley DJ and Vere-Jones D, An introduction to the theory of point processes. Vol. I, Probability and its Applications (New York). Springer-Verlag, New York, second edition, 2003, Elementary theory and methods. [Google Scholar]
  • [16].Burkholder Donald L., “Sharp inequalities for martingales and stochastic integrals,” in Colloque Paul Lévy sur les processus stochastiques, number 157–158 in Astérisque, pp. 75–94. Société mathématique de France, 1988. [Google Scholar]
  • [17].Rodrigo Bañuelos and Gang Wang, “Sharp inequalities for martingales with applications to the beurling-ahlfors and riesz transforms,” Duke Math. J, vol. 80, no. 3, pp. 575–600, 12 1995. [Google Scholar]
  • [18].Young LC, “An inequality of the Hölder type, connected with Stieltjes integration,” Acta Math, vol. 67, pp. 251–282, 1936. [Google Scholar]
  • [19].Garsia AM, Rodemich E, and Rumsey H Jr., “A real variable lemma and the continuity of paths of some Gaussian processes,” Indiana University Mathematics Journal, vol. 20, no. 6, pp. 565–578, 1970. [Google Scholar]
  • [20].Shevchenko G, “Fractional Brownian motion in a nutshell,” in International Journal of Modern Physics Conference Series, Jan. 2015, vol. 36 of International Journal of Modern Physics Conference Series, p. 1560002. [Google Scholar]

RESOURCES