Skip to main content
Springer logoLink to Springer
. 2018 Mar 17;60(9):1427–1458. doi: 10.1007/s10851-018-0806-0

Design and Processing of Invertible Orientation Scores of 3D Images

M H J Janssen 1, A J E M Janssen 2, E J Bekkers 1, J Oliván Bescós 3, R Duits 1,
PMCID: PMC6413631  PMID: 30956394

Abstract

The enhancement and detection of elongated structures in noisy image data are relevant for many biomedical imaging applications. To handle complex crossing structures in 2D images, 2D orientation scores U:R2×S1C were introduced, which already showed their use in a variety of applications. Here we extend this work to 3D orientation scores U:R3×S2C. First, we construct the orientation score from a given dataset, which is achieved by an invertible coherent state type of transform. For this transformation we introduce 3D versions of the 2D cake wavelets, which are complex wavelets that can simultaneously detect oriented structures and oriented edges. Here we introduce two types of cake wavelets: the first uses a discrete Fourier transform, and the second is designed in the 3D generalized Zernike basis, allowing us to calculate analytical expressions for the spatial filters. Second, we propose a nonlinear diffusion flow on the 3D roto-translation group: crossing-preserving coherence-enhancing diffusion via orientation scores (CEDOS). Finally, we show two applications of the orientation score transformation. In the first application we apply our CEDOS algorithm to real medical image data. In the second one we develop a new tubularity measure using 3D orientation scores and apply the tubularity measure to both artificial and real medical data.

Keywords: Orientation scores, 3D wavelet design, Zernike polynomials, Scale spaces on SE(3), Coherence-enhancing diffusion, Tubular structure detection, Steerable 3D wavelet

Introduction

The enhancement and detection of elongated structures are important in many biomedical image analysis applications. These tasks become problematic when multiple elongated structures cross or touch each other in the data. In these cases it is useful to work with multi-orientation representations of image data. Such multi-orientation representations can be made using various techniques, such as invertible orientation scores (which is obtained via a coherent state transform) [3, 5, 10, 30, 36, 42], continuous wavelet transforms [10, 28, 30, 64], orientation lifts [13, 71], or orientation channel representations [35]. Here we focus on constructing an invertible orientation score. In order to separate the crossing or touching structures (Fig. 1), we extend the domain of the data to include orientation. This is achieved by correlating our 3D data f:R3R with a set of oriented filters to construct a 3D orientation score U:R3×S2C. As the transformation between image and orientation score is stable, due to our design of anisotropic wavelets, we can robustly relate operators on the score to operators on images. To take advantage of the multi-orientation decomposition, we consider operators on orientation scores and process our data via orientation scores (Fig. 2).

Fig. 1.

Fig. 1

2D orientation score for an exemplary image. In the orientation score crossing structures are disentangled because the different structures have a different orientation

Fig. 2.

Fig. 2

A schematic view of image processing via invertible orientation scores

Regarding the invertibility of the transform from image to orientation score, we note that in comparison to continuous wavelet transforms (see, e.g., [4, 48, 50, 51] and many others) on the group of 3D rotations, translations and scalings, we use all scales simultaneously and exclude the scaling group from the wavelet transform and its adjoint, yielding a coherent state type of transform [2]. This makes it harder to design appropriate wavelets, but has the computational advantage of only needing one all-scale transformation.

The 2D orientation scores have already showed their use in a variety of applications. In [37, 64] the orientation scores were used to perform crossing-preserving coherence-enhancing diffusions. These diffusions greatly reduce the noise in the data, while preserving the elongated crossing structures. Next to these generic enhancement techniques, the orientation scores also showed their use in retinal vessel tracking [8, 10, 19], in vessel segmentation [70] and biomarker analysis [11, 60], where they were used to better handle crossing vessels. Here we aim to extend such techniques to 3D data.

To perform detection and enhancement operators on the orientation score, we first need to transform a given grayscale image or 3D dataset to an orientation score in an invertible way. In previous works various wavelets were introduced to perform a 2D orientation score transform. Some of these wavelets did not allow for an invertible transformation (e.g., Gabor wavelets [48]). A wavelet that allows an invertible transformation was proposed by Kalitzin et al. [46]. A generalization of these wavelets was found by Duits [25] who derived formal unitarity results and expressed the wavelets in a basis of eigenfunctions of the harmonic oscillator. This type of wavelet was also extended to 3D. This wavelet, however, has some unwanted properties such as poor spatial localization (oscillations) and the fact that the maximum of the wavelet does not lie at its center. In [25] a class of 2D cake wavelets was introduced that have a cake-piece-shaped form in the Fourier domain. The cake wavelets simultaneously detect oriented structures and oriented edges by constructing a complex orientation score U:R2×S1C. Because the family of rotated cake wavelets cover the full Fourier spectrum, invertibility is guaranteed.

In this article we propose a 3D version of the cake wavelets. A preliminary attempt to generalize these filters was done in [44], where the plate detectors in [25] were extended to complex-valued cake wavelets with a line detector in the real part. Compared to these previous works, the filters in this work are now exact until sampling in the Fourier domain. For these filters we have no analytical description in the spatial domain as filters are obtained via a discrete inverse Fourier transform (DFT). Therefore, we additionally consider expressing filters of this type in the 3D generalized Zernike basis. For this basis we have analytical expressions for the inverse Fourier transform, allowing us to find analytical expressions for the spatial filters. This has the additional advantage that they allow for an implementation with steerable filters, see App. A. These analytical expressions are then used to validate the filters obtained using the DFT method. We also show applications of these filters and the orientation score transformation in 3D vessel analysis. That is, we present crossing-preserving diffusions for denoising 3D rotational Xray of blood vessels in the abdomen and we present a tubularity measure via orientation scores and features based on this tubularity measure, which we apply to cone beam CT data of the brain. An overview of the applications is presented in Fig. 3. Regarding our nonlinear diffusions of 3D rotational Xray images via invertible orientation scores, we observe that complex geometric structures in the vasculature (involving multiple orientations) are better preserved than with nonlinear diffusion filtering directly in the image domain. This is in line with previous findings for nonlinear diffusion filtering of 2D images [37] and related works [54, 63, 66] that rely on other more specific orientation decompositions.

Fig. 3.

Fig. 3

Overview of applications of processing via orientation scores. Top: We reduce noise in real medical image data (3D rotational Xray) of the abdomen containing renal arteries by applying diffusions via 3D orientation scores. Bottom: Geometrical features can be directly extracted from our tubularity measure via 3D orientation scores. We apply this method to cone beam CT data of the brain

For the sake of general readability, we avoid Lie group theoretical notations, until Sect. 6.1 where it is strictly needed. Let us nevertheless mention that our work fits in a larger Lie group theoretical framework, see, for example, [3, 7, 25, 26, 39] that has many applications in image processing. Besides the special cases of the Heisenberg group [6, 31, 57], the 2D Euclidean motion group [5, 9, 21, 22, 30, 53, 59], the similitude group [4, 56, 64] and the 3D rotation group [52], we now consider invertible orientation scores on the 3D Euclidean motion group (in which the coupled space of positions and orientations is embedded). The invertible orientation scores relate to coherent states from physics [42] for n=3, with a specific (semi-)analytic design for our image processing purposes.

Contributions of the Article

The main contributions per section of the article are:

  • In Sect. 2 we give an overview of the discrete and continuous 3D orientation score transformation. Additionally, we present a transformation which is split in low and high frequencies and quantify the stability of the transformation in Lemma 1. item In Sect. 3 we present the cake wavelets obtained using the DFT method and give an efficient implementation using spherical harmonics which is summarized in Result 1. Furthermore, we analyze the stability of the transformation for these filters (Proposition 1).

  • In Sect. 4 we present the analytical versions of the cake wavelets obtained by expansion in the Zernike basisfollowed by a continuous Fourier transform which is summarized in Result 2.

  • In Sect. 5 we compare the two types of filters and show that the DFT filters approximate their analytical counterparts well.

  • In Sect. 6 we show two applications of the orientation score transformation:
    1. We propose an extension of coherence-enhancing diffusion via our invertible orientation scores of 3D images. Compared to the original idea of coherence-enhancing diffusion acting directly on image data [16, 17, 69], there is the advantage of preserving crossings. Here we applied our method to real medical image data (3D rotational Xray) of the abdomen containing renal arteries. We show quantitatively that our method effectively reduces noise (quantified using contrast-to-noise ratios (CNR)) while preserving the complex vessel geometry and the vessel widths. Furthermore, qualitative assessment indicates that our denoising method is very useful as preprocessing for 3D visualization (volume rendering).
    2. We develop a new tubularity measure in 3D orientation score data. This extends previous work on tubularity measures using 2D orientation scores [19] [9, Ch. 12] to 3D data. We show qualitatively that our measure gives sharp responses at vessel centerlines and show its use for radius extraction and complex vessel segmentation in cone beam CT data of the brain.

Outline of the Article

To summarize, we give a global outline of the article: First, we discuss the theory of invertible orientation score transforms in Sect. 2. Then we construct 3D cake wavelets and give a new efficient implementation using spherical harmonics in Sect. 3, followed by their analytical counterpart expressed in the generalized Zernike basis in Sect. 4. Then we compare the two types of filters and validate the invertibility of the orientation score transformation in Sect. 5. Finally, we address two application areas for 3D orientation scores in Sect. 6 and show results and practical benefits for both of them. In the first application (Sect. 6.1), we present a natural extension of the crossing-preserving coherence-enhancing diffusion on invertible orientation scores (CEDOS) [37] to the 3D setting. In the second application (Sect. 6.2) we use the orientation score to define a tubularity measure and show experiments applying the tubularity measure to both synthetic data and brain CT data. Both application sections start with a treatment of related methods.

Invertible Orientation Scores

Continuous Orientation Score Transform

Throughout this article we use the following definition of the Fourier transform on R3:

f^(ω)=(Ff)(ω)=R3e-iω·xf(x)dx. 1

An invertible orientation score Wψ[f]:R3×S2C is constructed from a given ball-limited 3D dataset

fL2ϱ(R3)={fL2(R3)|supp(Ff)Bϱ}, 2

with ball Bϱ={ωR3|ω<ϱ} of radius ϱ>0, by correlation with an anisotropic kernel

(Wψ[f])(x,n)=(ψn¯f)(x)=R3ψn(x-x)¯f(x)dx. 3

This is illustrated in Fig. 4. Here ψL2(R3)L1(R3) is a wavelet aligned with and rotationally symmetric around the z-axis, and ψnL2(R3) the rotated wavelet aligned with n given by

ψn(x)=ψ(RnTx). 4

Here RnSO(3) is any 3D rotation which rotates the z-axis onto n where the specific choice of rotation does not matter because of the rotational symmetry of ψ. The overline denotes a complex conjugate. The exact reconstruction formula [25] for this transformation is given by

f(x)=(Wψ-1[Wψ[f]])(x)=F-1Mψ-1Fx~S2(ψˇnWψ[f](·,n))(x~)dσ(n)(x), 5

with ψˇn(x)=ψn(-x). The function Mψ:R3R+ is given by

Mψ(ω)=S2ψ^n(ω)2dσ(n), 6

and vanishes at , where the circumflex (^) again denotes Fourier transformation. Due to our restriction to ball-limited data (2), this does not cause problems in reconstruction (5). The function Mψ quantifies the stability of the inverse transformation [25], since Mψ(ω) specifies how well frequency component ω is preserved by the cascade of construction and reconstruction when Mψ-1 would not be included in Eq. (5). An exact reconstruction is possible as long as

M>0,δ>0ωBϱ:0<δMψ(ω)M<. 7

In practice it is best to aim for Mψ(ω)1, in view of the condition number of transformation Wψ:L2ϱ(R3)L2ϱ(R3×S2) given by:

cond(Wψ)=WψWψ-1=Mδ, 8

where M and δ are assumed to be tight bounds in (7). In the codomain spatial frequencies are again limited to the ball:

L2ϱ(R3×S2)={UL2(R3×S2)|nS2U(·,n)L2ϱ(R3)}. 9

Also, in the case we have Mψ(ω)=1 for ωBϱ we have L2-norm preservation

fL2(R3)2=WψfL2(R3×S2)2,for allfL2ϱ(R3), 10

and reconstruction Eq. (5) simplifies to

f(x)=S2(ψˇnWψ[f](·,n))(x)dσ(n). 11

We can further simplify the reconstruction for wavelets for which the following additional property holds:

Nψ(ω)=S2ψ^n(ω)dσ(n)1. 12

In that case the reconstruction formula is approximately an integration over orientations only:

f(x)S2Wψ[f](x,n)dσ(n). 13

For the reconstruction by integration over angles only we can analyze the stability via the condition number of the mapping that maps an image fL2ϱ(R3) to an orientation integrated score

Aψ(f)=S2Wψf(·,n)dσ(n). 14

Its condition number is given by cond(Aψ)=maxωBϱNψ(ω)minωBϱNψ(ω).

Fig. 4.

Fig. 4

Construction of a 3D orientation score. Top: The data f are correlated with an oriented filter ψex to detect structures aligned with the filter orientation ex. Bottom left: This is repeated for a discrete set of filters with different orientations. Bottom right: The collection of 3D datasets constructed by correlation with the different filters is an orientation score and is visualized by placing a 3D dataset on a number of orientations

In practice, we always use this last reconstruction because practical experiments show that performing an additional convolution with the wavelet as done in reconstruction (5) after processing the score can lead to artifacts. It is, however, important to also consider the reconstruction (5) and Mψ because it is used to quantify the stability and norm preservation of the transformation from image to orientation score.

The fact that we use reconstruction by integration while still taking into account norm preservation by controlling Mψ leads to restrictions on our wavelets which are captured in the following definition:

Definition 1

(Proper wavelet) Let us set a priori bounds1 δ,M>0,0<ε1. Furthermore, let ϱ be an a priori maximum frequency of our ball-limited image data. Then, a wavelet ψL2(R3)L1(R3) is called a proper wavelet if

1.)α[0,2π):ψ(Rez,α-1x)=ψ(x), 15
2.)ωBϱ:δMψ(ω)M, 16

where Rez,αSO(3) is the 3D rotation around axis ez over angle α.

If, moreover, one has

3.)12ϱ<ϱ0<ϱωBϱ0:Nψ(ω)[1-ε,1+ε], 17

then we speak of a proper wavelet with fast reconstruction property, cf. (13).

Remark 1

The 1st condition (symmetry around the z-axis) allows for an appropriate definition of an orientation score rather than a rotation score. The 2nd condition ensures invertibility and stability of the (inverse) orientation score transform. The 3rd condition allows us to use the approximate reconstruction by integration over angles only.

Remark 2

Because of finite sampling in practice, the constraint to ball-limited functions is reasonable. The constraint is not a necessary one when one relies on distributional transforms [10, App. B], but we avoid such technicalities here.

Low-Frequency Components

In practice we are not interested in the zero and lowest frequency components since they represent average value and global variations which appear at scales much larger than the structures of interest. We need, however, to store these data for reconstruction. Therefore, we perform an additional splitting of our wavelets into two parts

ψ=ψ0+ψ1,withψ^0=G^sρψ^,ψ^1=(1-G^sρ)ψ^, 18

with Gaussian window in the Fourier domain given by

G^sρ(ω)=e-sρω2, 19

After splitting, ψ0 contains the average and low-frequency components and ψ1 the higher frequencies relevant for further processing. In continuous wavelet theory it is also common to separately store very low-frequency components separately, see, e.g., [51, 65]. In this case we construct two scores: one for the high-frequency components

(Wψ1[f])(x,n)=(ψ1,n¯f)(x), 20

and one for the low-frequency components

(Wψ0[f])(x,n)=(ψ0,n¯f)(x). 21

Here we again have ψi,n(x)=ψi(RnTx), as in Eq. (4). The vector transformation is then defined as

W_ψ_[f]=(Wψ0[f],Wψ1[f]). 22

For this transformation we have the exact reconstruction formula

f(x)=(W_ψ_-1W_ψ_f)(x)=F-1[Mψ_-1F[x~S2(ψˇ1,nWψ1[f](·,n))(x~)+(ψˇ0,nWψ0[f](·,n))(x~)dσ(n)]](x) 23

with

Mψ_(ω)=S2ψ^0,n(ω)2+ψ^1,n(ω)2dσ(n). 24

Again, Mψ_ quantifies the stability of the transformation. The next lemma shows us that the stability of the transformation is maintained after performing the additional splitting.

Lemma 1

Let ψL2(R3)L1(R3) such that Eq. (7) holds, δ=minωBϱMψ(ω) and M=maxωBϱMψ(ω). Then the condition number of Wψ:L2ϱ(R3)L2ϱ(R3×S2) is given by

|cond(Wψ)|2=Wψ2Wψ-12=Mδ. 25

The condition number of W_ψ_:L2ϱ(R3)L_2ϱ(R3×S2) obtained from Wψ by performing an additional splitting in low- and high-frequency components is given by

|cond(W_ψ_)|2=W_ψ_2W_ψ_-122Mδ, 26

thereby guaranteeing that stability is maintained after performing the splitting.

Proof

First, we find the condition number of Wψ:

|cond(Wψ)|2=supfL2ϱ(R3)fL22WψfL22·supfL2ϱ(R3)WψfL22fL22. 27

For the first factor in Eq. (27) we find

supfL2ϱ(R3)fL22WψfL22=supfL2ϱ(R3)FfL22FWψfL22=supfL2ϱ(R3)R3|f^(ω)|2dωS2R3|ψ^n(ω)|2|f^(ω)|2dωdσ(n)=supfL2ϱ(R3)R3|f^(ω)|2dωR3Mψ(ω)|f^(ω)|2dω=maxωBϱ1Mψ(ω). 28

where in the last step the supremum is attained by a sequence of images whose Fourier transform concentrates around the maximum of Mψ. Similarly, we get maxωBϱMψ(ω) for the second factor in Eq. (27). Then we obtain

cond(Wψ)=maxωBϱ1Mψ(ω)·maxωBϱMψ(ω)=Mδ. 29

Similarly the condition number of W_ψ_ is given by

cond(W_ψ_)=maxωBϱ1Mψ_(ω)·maxωBϱMψ_(ω). 30

Next we express Mψ_ in Mψ of the original wavelet as

Mψ_(ω)=S2ψ^0,n(ω)2+ψ^1,n(ω)2dσ(n)=S2ψ^0,n(ω)+ψ^1,n(ω)2dσ(n)-S22Reψ^0,n(ω)ψ^1,n(ω)¯dσ(n)=Mψ(ω)-I(ω). 31

So it remains to quantify I(ω). For a wavelet splitting according to (18) we have

I(ω)=S22Reψ^0,n(ω)ψ^1,n(ω)¯dσ(n)=S22ReG^sρ(ω)ψ^n(ω)(1-G^sρ(ω))ψ^n(ω)¯dσ(n)=2(G^sρ(ω)(1-G^sρ(ω)))Mψ(ω). 32

Hence

Mψ_(ω)=(1-2(G^sρ(ω)(1-G^sρ(ω))))Mψ(ω). 33

And since 121-2x(1-x))1 for 0x1 we have for Mψ satisfying (7) the following bounds on Mψ_:

0<δ/2Mψ_(ω)M<,for allω=Bϱ, 34

thereby guaranteeing stability after the splitting (18). As we cannot guarantee that δ/2 and M are tight bounds in Eq. (34), combining it with (29) will only give us an upper bound for the condition number in Eq. (26).

For this vector transformation we can also use the approximate reconstruction by integration (for Nψ1) over orientations. Thus we have

f(x)S2Wψ[f](x,n)dσ(n)=S2Wψ1[f](x,n)dσ(n)+S2Wψ0[f](x,n)dσ(n)Lψ0[f](x). 35

As said we are only interested in processing of Wψ1[f] and not in processing of Wψ0[f], and so we directly calculate Lψ0[f] via

Lψ0[f](x)=(ϕ0¯f)(x),withϕ0=S2ψ0,ndσ(n). 36

For a design with Nψ(ω)=1 for all ωBϱ, we have ϕ^0=G^sρ and so

ϕ0(x)=Gsρ(x)=1(4πsρ)3/2e-x24sρ. 37

Then Eq. (35) becomes

f(x)S2Wψ1[f](x,n)dσ(n)+(Gsρf)(x). 38

Discrete Orientation Score Transform

In the previous section, we considered a continuous orientation score transformation. In practice, we have only a finite number of orientations. To determine this discrete set of orientations we uniformly sample the sphere using a simple2 electrostatic repulsion model [18].

Assume we have a number No of orientations V={n1,n2,,nNo}S2, and define the discrete invertible orientation score Wψd[f]:R3×VC by

(Wψd[f])(x,ni)=(ψni¯f)(x). 39

The exact reconstruction formula is in the discrete setting given by

f(x)=((Wψd)-1[Wψd[f]])(x)=F-1[(Mψd)-1F[x~i=1No(ψˇniWψd[f](·,ni))(x~)Δi]](x), 40

with Δi the discrete spherical area measure (i=1NoΔi=4π) which for reasonably uniform spherical sampling can be approximated by Δi4πNo, and

Mψd(ω)=i=1Noψ^ni(ω)2Δi. 41

For spherical samplings that are constructed by triangularization (such as tessellations of the icosahedron), one can rely on surface areas of spherical triangles to compute Δi more accurately. See for example [29, Eq. 83]).

Again, an exact reconstruction is possible iff 0<δMψd(ω)M< and we have norm preservation when Mψd=1. Again, for the wavelets for which

Nψd=i=1Noψ^ni(ω)Δi1, 42

the image reconstruction can be simplified by a summation over orientations:

f(x)i=1NoWψd[f](x,ni)Δi. 43

For this reconstruction by summation we can analyze the stability via the condition number of the mapping that maps an image fL2ϱ(R3) to an orientation integrated score

Aψd(f)=i=1Noψ¯nifΔi. 44

This transformation has condition number cond(Aψd)=maxωBϱNψd(ω)minωBϱNψd(ω).

Similar to Definition 1 for the continuous case, the reconstruction properties of a set of filters is captured in the following definition:

Definition 2

(Proper wavelet set) Let us again set a priori bounds δ,M>0,0<ε1. Let ϱ be an a priori maximum frequency of our ball-limited image data. Then, a set of wavelets {ψniL2ϱ(R3)L1(R3)|i=1,,No}, with a reasonable uniform spherical sampling (Δi4πNo), constructed as rotated versions of ψ is called a proper wavelet set if

1.)α[0,2π):ψ(Rez,α-1x)=ψ(x), 45
2.)ωBϱ:δMψd(ω)M, 46

where Rez,αSO(3) is a 3D rotation around axis ez over angle α.

If, moreover, one has

3.)12ϱ<ϱ0<ϱωBϱ0:Nψd(ω)[1-ε,1+ε], 47

then we speak of a proper wavelet with fast reconstruction property, cf. (43).

Low-Frequency Components

For the discrete transformation we will also perform a splitting in low- and high-frequency components as explained in Sect. 2.1.1. The reconstruction formula by summation in Eq. (43) is now given by

f(x)i=1NoWψ1d[f](x,ni)Δi+(Gsρf)(x). 48

Steerable Orientation Score Transform

Throughout this article we shall rely on a spherical harmonic decomposition of the angular part of proper wavelets in the spatial and the Fourier domain. This has the benefit that one can obtain steerable [36, 38, 61] implementations of orientation scores, where rotations of the wavelets are obtained via linear combination of the basis functions. As such, computations are exact and no interpolation (because of rotations) takes place. Details are provided in Appendix A.

Wavelet Design Using a DFT

A class of 2D cake wavelets, see [10, 27, 37], was used for the 2D orientation score transformation. We now present 3D versions of these cake wavelets. Thanks to the splitting in Sect. 2.1.1 we no longer need the extra spatial window used there. Our 3D transformation using the 3D cake wavelets should fulfill a set of requirements, compare [37]:

  1. The orientation score should be constructed for a finite number (No) of orientations.

  2. The transformation should be invertible, and reconstruction should be achieved by summation. Therefore, we aim for Nψd1. Additionally, to guarantee all frequencies are transferred equally to the orientation score domain we aim for Mψd1. The set should be a proper wavelet set with fast reconstruction property (Definition 2).

  3. The kernel should be strongly directional.

  4. The kernel should be separable in spherical coordinates in the Fourier domain, more explicitly (Fψ)(ω)=g(ρ)h(ϑ,φ), with
    ω=(ωx,ωy,ωz)=(ρsinϑcosφ,ρsinϑsinφ,ρcosϑ). 49
    Because by definition the wavelet ψ has rotational symmetry around the z-axis we have Inline graphic.
  5. The kernel should be localized in the spatial domain, since we want to pick up local oriented structures.

  6. The real part of the kernel should detect oriented structures, and the imaginary part should detect oriented edges. The constructed orientation score is therefore a complex-valued orientation score, as the wavelets are complex-valued. For an intuitive preview, see the boxes in Fig. 7.

Fig. 7.

Fig. 7

Cake Wavelets. Top: 2D cake wavelets. From left to right: Illustration of the Fourier domain coverage, the wavelet in the Fourier domain and the real and imaginary part of the wavelet in the spatial domain [10]. Bottom: 3D cake wavelets. Overview of the transformations used to construct the wavelets from a given orientation distribution. Upper part: The wavelet according to Eq. (53). Lower part: The wavelet according to Eq. (54). IFT: Inverse Fourier Transform. Parameters used: so=12(0.25)2,γ=0.85 and evaluated on a grid of 81×81×81 pixels

Construction of Line and Edge Detectors

We now discuss the procedure used to make 3D cake wavelets before splitting in low and high frequencies according to (18) in Sect. 2.1.1 takes place. Following requirement 4 we only consider polar separable wavelets in the Fourier domain, so that Inline graphic. To satisfy requirement 2 we should choose radial function g(ρ)=1 for ρ[0,ϱ]. In practice, this function should go to 0 when ρ tends to the Nyquist frequency ρN to avoid long spatial oscillations. For the radial function g(ρ) we use,

g(ρ)=121-erfρ-ϱσerf, 50

with erf(z)=2π0ze-x2dx, which is approximately equal to one for largest part of the domain and then smoothly goes to 0 when approaching the Nyquist frequency. We fix the inflection point of this function g and set the fundamental parameter for ball limitedness to

ϱ=γρN, 51

with 0γ<1. The steepness of the decay when approaching ρN is controlled by the parameter σerf which we by default set to σerf=13(ρN-ϱ). The additional splitting in low and high frequencies according to Sect. 2.1.1 effectively causes a splitting of the radial function, see Fig. 5.

Fig. 5.

Fig. 5

Radial part g of ψ^, see Eq. (50) and radial parts g0 and g1 of ψ^0 and ψ^1 after splitting according to Sect. 2.1.1. The parameter γ controls the inflection point of the error function, here γ=0.8. The steepness of the decay when approaching ρN is controlled by the parameter σerf with default value σerf=13(ρN-ϱ). At what frequency the splitting of ψ^ in ψ^0 and ψ^1 is done is controlled by parameter σρ=2sρ, see Eq. (18)

In practice the frequencies in our data are limited by the Nyquist frequency (we have ϱρN), and because radial function g causes Mψd to become really small close to the Nyquist frequency, reconstruction Eq. (40) becomes unstable. We solve this by using approximate reconstruction Eq. (13). Alternatively, one could replace Mψd by max(Mψd,ϵ) in Eq. (5), with ϵ small. Both make the reconstruction stable at the cost of not completely reconstructing the highest frequencies which causes a small amount of blurring.

We now need to find an appropriate angular part Inline graphic for the cake wavelets. First, we specify an orientation distribution A:S2R+, which determines what orientations the wavelet should measure. To satisfy requirement 3 this function should be a localized spherical window, for which we propose the spherical diffusion kernel [20]:

A(n(ϑ,φ))=GsoS2(n(ϑ,φ)), 52

with so>0 and n(ϑ,φ)=(sinϑcosφ,sinϑsinφ,cosϑ). The parameter so determines the trade-off between requirements 2 and 3 listed in the beginning of Sect. 3, where higher values give a more uniform Mψd at the cost of less directionality.

First consider setting h=A so that ψ has compact support within a convex cone in the Fourier domain. The real part of the corresponding wavelet would, however, be a plate detector and not a line detector (Fig. 6). The imaginary part is already an oriented edge detector,3 and so we set

graphic file with name 10851_2018_806_Equ53_HTML.gif 53

where the real part of the earlier found wavelet vanishes by anti-symmetrization of the orientation distribution A while the imaginary part is unaffected. Note that the right-hand side of (52) and (53) is the same for all values of φ. As to the construction of hRe, there is the general observation that we detect a structure that is perpendicular to the shape in the Fourier domain, so for line detection we should aim for a plane detector in the Fourier domain. To achieve this we apply the Funk transform to A, and we define

hRe(ϑ,φ)=FA(n(ϑ,φ))=12πSp(n(ϑ,φ))A(n)ds(n), 54

where integration is performed over Sp(n) denoting the great circle perpendicular to n. This transformation preserves the symmetry of A, so we have Inline graphic. Thus, we finally set

graphic file with name 10851_2018_806_Equ55_HTML.gif 55

For an overview of the transformations, see Fig. 7.

Fig. 6.

Fig. 6

When directly setting orientation distribution A of Eq. (35) as angular part of the wavelet h we construct plate detectors. From left to right: Orientation distribution A, wavelet in the Fourier domain, the plate detector (real part) and the edge detector (imaginary part). Orange: Positive iso-contour. Blue: Negative iso-contour. Parameters used: so=12(0.25)2,σerf=3,γ=0.85 and evaluated on a grid of 51×51×51 pixels (Color figure online)

As discussed before, the additional splitting in low and high frequencies as described in Sect. 2.1.1 effectively causes a splitting in the radial function. How this affects the coverage of the Fourier domain is visualized in Fig. 8.

Fig. 8.

Fig. 8

Coverage of the Fourier domain before and after splitting according to Sect. 2.1.1. Left: The different wavelets cover the Fourier domain. The “sharp” parts when the cones reach the center, however, cause the filter to be non-localized, which was solved in earlier works by applying a spatial window after filter construction. Right: By splitting the filter in lower and higher frequencies we solve this problem. In the figure we show g(ρ)A(n(ϑ,φ)) for the different filters, before applying the Funk transform to the orientation distribution A

Efficient Implementation Via Spherical Harmonics

In Sect. 3.1 we defined the real part and the imaginary part of the wavelets in terms of a given orientation distribution. In order to efficiently implement the various transformations (e.g., Funk transform) and to create the various rotated versions of the wavelet, we express our orientation distribution A in a spherical harmonic basis {Ylm} up to order L:

A(n(ϑ,φ))=l=0Lm=-llalmYlm(ϑ,φ),LN. 56

The spherical harmonics are given by

Ylm(ϑ,φ)=ϵ2l+14π(l-|m|)!(l+|m|)!eimφPl|m|(cosϑ), 57

where Plm is the associated Legendre function, ϵ=(-1)m for m<0 and ϵ=1 for m>0 and with integer l0 and integer m satisfying -lml. For the diffusion kernel, which has symmetry around the z-axis we only need the spherical harmonics with m=0, and we have the coefficients [20]:

alm=0m0,2l+14πe-l(l+1)som=0, 58

and Eq. (56) reduces to

A(n(ϑ,φ))=l=0Lal0Yl0(ϑ,φ). 59

Funk Transform

According to [23], the Funk transform of a spherical harmonic equals

FYlm(ϑ,φ)=12πSp(n(ϑ,φ))Ylm(n)ds(n)=Pl(0)Ylm(ϑ,φ), 60

with Pl(0) the Legendre polynomial of degree l evaluated at 0. We can therefore apply the Funk transform to a function expressed in a spherical harmonic basis by a simple transformation of the coefficients almPl(0)alm.

Anti-symmetrization

We have Ylm(π-ϑ,φ+π)=(-1)lYlm(ϑ,φ). We therefore anti-symmetrize the orientation distribution, see Eq. (53), via alm(1-(-1)l)2alm.

Making Rotated Wavelets

To make the rotated versions ψn of wavelet ψ, we have to find hn in ψ^n(ω)=g(ρ)hn(ϑ,φ). To achieve this we use the steerability of the spherical harmonic basis. Spherical harmonics rotate according to the irreducible representations of the SO(3) group Dm,ml(γ,β,α) (Wigner-D functions [41]):

RRγ,β,αYlm(ϑ,φ)=m=-llDm,ml(γ,β,α)Ylm(ϑ,φ). 61

Here α,β and γ denote the Euler angles with counterclockwise rotations, where we rely on the convention Rγ,β,α=Rez,γRey,βRez,α. This gives

hn(ϑ,φ)=RRγ,β,αh(ϑ,φ)=l=0Lm=-llm=-llclmDm,ml(γ,β,α)Ylm(ϑ,φ), 62

where clm are the coefficients of h given by

clm=Pl(0)alm+(1-(-1)l)2alm. 63

Because both anti-symmetrization and Funk transform preserve the rotational symmetry of A, we have h(ϑ,φ)=l=0Lcl0Yl0(ϑ,φ), and Eq. (62) reduces to

hn(ϑ,φ)=l=0Lm=-llcl0D0,ml(γ,β,0)Ylm(ϑ,φ). 64

The filters from this section are summarized in the following result:

Result 1

Let A:S2R+ be a function supported mainly in a sharp convex cone around the z-axis and symmetrically around the z-axis and g as radial function of Eq. (50). Then A provides our wavelet ψ^ in the Fourier domain via

ψ^(ω)=g(ρ)FA(nω)+A(nω)-A(-nω), 65

with ω=ρnω=ρn(ϑ,φ). The real part of ψ is a tube detector given by

Re(ψ)=F-1ωg(ρ)(FA)(nω). 66

The imaginary part of ψn is an edge detector given by

Im(ψ)=1iF-1ωg(ρ)A(nω)-A(-nω). 67

When expanding the angular part in spherical harmonics up to order L and choosing A=GsoS2:

A(n(ϑ,φ))=l=0Lal0Yl0(ϑ,φ),al0=2l+14πe-l(l+1)so, 68

we have the following wavelet in the Fourier domain

ψ^(ω)=g(ρ)l=0Lcl0Yl0(ϑ,φ), 69

and the coefficients of A and ψ^ relate via

cl0=Pl(0)+(1-(-1)l)2al0. 70

We obtain rotated versions of our filter via

ψ^n(ω)=g(ρ)l=0Lm=-llcl0D0,ml(γ,β,0)Ylm(ϑ,φ), 71

with n=n(β,γ).

As we do not have analytical expressions for the spatial wavelets ψn, we sample the filter in the Fourier domain using Eq. (71) and apply a DFT afterward. The wavelet ψ is a proper wavelet with fast reconstruction property (Definition 1).

Remark 3

The heat kernel on S2 is given by

GsoS2(n(ϑ,φ))=l=0al0Yl0(ϑ,φ)=l=02l+14πe-so(l+1)lYl0(ϑ,φ), 72

where we recall Eq. (68). Because of the exponential decay with respect to l, we can describe the diffusion kernel well with the first few coefficients. In all experiments we truncate at smallest L such that aL0/a00<10-3 (e.g., L=21 for so=12(0.25)2).

Stability of the Discrete Transformation with Fast Reconstruction for Filters of Result 1

To make a fast reconstruction by summation possible (requirement 2), we need a proper wavelet set with the fast reconstruction property (Definition 2) with Nψd1. We now focus on finding bounds for Nψd such that we can choose our parameters in a deliberate way.

Proposition 1

Let {ψn(βi,γi)|i=1No} be a set of wavelets constructed via the procedure in Result 1. Then we have bounds on Nψd given by

1-l=1Ldl2l+14πNψd(ω)1+l=1Ldl2l+14π,for allωBρ0 73

with dl=(dlm)m=-ll, dlm=i=1Nocl0·Δi·D0,ml(0,βi,γi) and here the norm is the 2-norm on C2l+1.

Proof

First we expand function Nψd in spherical harmonics:

Nψd(ω)=i=1NoF[ψni](ω)Δi=g(ρ)i=1Nohni(ϑ,φ)Δi=g(ρ)l=0Lm=-lli=1Nocl0D0,ml(0,βi,γi)ΔidlmYlm(ϑ,φ)=g(ρ)l=0Lm=-lldlmYlm(ϑ,φ) 74

We have g(ρ)=1 for ρ=ωρ0, but we still need to quantify the angular part. We define YlN=(Yl-l,Yl-l+1,,Yll-1,Yll), so that

l=0Lm=-lldlmYlm(ϑ,φ)=l=0Ldl·Yl(ϑ,φ)=Y00(ϑ,φ)d00+l=1Ldl·Yl(ϑ,φ)=1+l=1Ldl·Yl(ϑ,φ) 75

This varying component should remain small. We use the Cauchy-Schwarz inequality for each order l:

l=1Ldl·Yl(ϑ,φ)l=1Ldl·Yl(ϑ,φ)l=1LdlYl(ϑ,φ)=l=1Ldl2l+14π, 76

from which (73) follows.

See Fig. 9 for visual inspection of bounds of Mψd and Nψd, and numerical results for the bounds of Nψd.

Fig. 9.

Fig. 9

Inspection of the stability of the transformation for different values of so given an orientation distribution A=GsoS2 and for No=42. Left: Spherical plot of A and the angular part of polar separable function Nψd and Mψd. Orientation coverage is more uniform as the plots forNψd and Mψd look more like a ball. Right: The upper and lower bounds of Nψd. Comparison of the bounds according to Eq. (73) (filled blue line) and numerical results (dashed blue line) of the bounds by a very fine sampling of the sphere (500 orientations). Furthermore, we show 1+ϵ and 1-ϵ (orange dashed lines) for ϵ=0.05 (Color figure online)

Corollary 1

Given our analytical bounds (73) from Proposition 1 and No=42, we can guarantee that our set of wavelets from Result 1 is a proper wavelet set with fast reconstruction property according to Definition 2 with ϵ=0.05 when choosing parameter s00.04. In practice we have a proper wavelet set with fast reconstruction property already for smaller values of so (see Fig. 9).

Wavelet Design with Continuous Fourier Transform and Analytical Description in the Spatial Domain

In the previous section we described wavelets which were analytical in the Fourier domain and were sampled and inverse discrete Fourier transformed to find the wavelets in the spatial domain.

To get more control on the wavelet properties in both the spatial and Fourier domain, it would be convenient to have an analytical description of the wavelets in both domains. This could be achieved by expressing the wavelets in a basis for which we have analytical expressions for the Fourier transform. We will now discuss 2 such options for the basis and describe filters expressed in them.

A Review on Expansion in the Harmonic Oscillator Basis

The first basis in which we could expand our wavelets are the eigenfunctions of the harmonic oscillator H=x2-Δ, which was also studied in [25, 67]. We will quickly review this work and show the problems which were encountered when using this basis, before moving onto an alternative basis in the next section which aims to solve these problems.

When using the eigenfunctions of the harmonic oscillator as a basis, the idea is that operator H and the Fourier transform commute (FH=HF) and eigenfunctions of H are also eigenfunctions of F. We then expand our wavelets in these eigenfunctions restricting ourselves to eigenfunctions which are symmetric around the z-axis: the spherical harmonics with m=0. The wavelet is then given by

ψ(x)=n=0l=0Lαlngnl(r)Yl0(θ,ϕ),ψ^(ω)=n=0l=0Lαln(-1)n+lilgnl(ρ)Yl0(ϑ,φ), 77

with Ylm the spherical harmonics, (r,θ,ϕ) and (ρ,ϑ,φ) spherical coordinates for x and ω, respectively, i.e.,

x=(rsinθcosϕ,rsinθsinϕ,rcosθ),ω=(ρsinϑcosφ,ρsinϑsinφ,ρcosϑ), 78

and gnl given by

gnl(ρ)=1ρEnl+12(ρ),Enν(ρ)=2(n!)Γ(n+ν+1)12ρν+12e-ρ22Ln(ν)(ρ2), 79

where Ln(ν)(ρ) is the generalized Laguerre polynomial. We then choose the case with least radial oscillations αln=αlδn0. If we then choose

αl=Γl+32Γ(l+1), 80

we have that Mψ(ω) approximates 1 for all ωR3 in the Fourier domain as L and we get the following wavelet4 [27]:

ψH(x)=l=0L1l!rle-r22Yl0(θ,ϕ). 81

For this wavelet we have an analytical description in both spatial and Fourier domain.

Expansion in the Zernike Basis

The wavelets from the previous subsection have some unwanted properties such as poor spatial localization (long oscillations) and the fact that the wavelets maximum does not lie at the wavelets center, see Fig. 10. A possible explanation is that the basis used is orthogonal on the full L2(R3) space and not limited to the ball in the Fourier domain, and truncation of this basis at the Nyquist frequency could lead to oscillations. An alternative basis for the unit ball is the Zernike basis which we can scale to be a basis in the Fourier domain for ball-limited images fL2ϱ(R3), recall Eq. (2).

Fig. 10.

Fig. 10

Wavelet expanded in the harmonic oscillator basis according to Eq. (81) for L=15. Left: 3D visualization showing one negative (blue) and one positive (orange) iso-contour. Right: Cross section of the wavelet at x=0 (Color figure online)

The 2D Zernike basis is often used in applications as optics, lithography and acoustics [1, 12, 15], since efficient recursions can be used for calculating the basis functions and analytic formulas exist for many transformations among which the Fourier transform. The basis is therefore highly suitable for problems such as aberration retrieval where a wave function in the Fourier domain should be estimated from measurements in the image domain [24].

Orthogonal polynomials of several variables on the unit ball were considered by Louis [49] in the context of inversion of the Radon transform for tomographic applications. For a modern treatment of orthogonal polynomials on the unit ball, see [33, Ch. 4]. Here we will use the generalized Zernike functions [43], which vanish to a prescribed degree at the boundary of the unit ball, with explicit expansion results for particular functions supported by the unit ball. Since this basis is orthogonal on the unit ball and has explicit results for the Fourier transform it is highly suitable for the application in mind. Derivations for the results used in the following section can be found in [43].

The 3D Generalized Zernike Basis

The generalized Zernike functions are given by

Zn,lm,α(ω)=Rnl,α(ρ)Ylm(ϑ,φ), 82

with spherical coordinates

ω=ρn(ϑ,φ), 83

integer n,l0 such that n=l+2p, integer p0 and m=-l,-l+1,,l and α>0. The angular part is the spherical harmonic function and the radial part is given by

Rnl,α(ρ)=ρl(1-ρ2)αPp=n-l2α,l+12(2ρ2-1), 84

where Pp(α,l+12) denotes the Jacobi polynomial. The generalized Zernike functions are orthogonal on the unit ball

ω1Zn1,l1m1,α(ω)Zn2,l2m2,α(ω)¯dω(1-ρ2)α=Nn,lαδn1,n2δm1,m2δl1,l2, 85

with δ the Kronecker delta and with normalization factor

Nn,lα=(p+1)αp+l+32α12n+α+32, 86

in which (x)α=Γ(x+α)Γ(x) is the (generalized) Pochhammer symbol.

Fourier Transform

The inverse Fourier transform of the generalized Zernike function

(F-1Zn,lm,α)(2πx)=ω1e2πi(ω·x)Rnl,α(ρ)Ylm(ϑ,φ)dω 87

is given by

(F-1Zn,lm,α)(2πx)=4πilSn,lα(2πr)Ylm(θ,ϕ), 88

with x=rn(θ,ϕ) and

Sn,lα(q)=01Rnl,α(ρ)jl(qρ)ρ2dρ=2α(-1)p(p+1)απ2qJn+α+32(q)qα+1ifq>0,πΓ(1+α)4Γ(52+α)δn,0ifq=0. 89

Here Ja and ja are the Bessel functions and spherical Bessel functions [55]. For integer α, the expression in Eq. (89) for q>0 reduces to

2α(-1)p(p+1)αjn+α+1(q)qα+1. 90

Expansion of Separable Functions

An additional constraint for the wavelets is that they should be separable in the Fourier domain, i.e., (Fψ)(ω)=F(ω)=A(ϑ,φ)B(ρ). When expanding such a function in the generalized Zernike basis,

F(ω)=n,l,mcn,lm,α(F)Zn,lm,α(ω), 91

we can split the coefficients in radial coefficients and angular coefficients

cn,lm,α(F)=1Nn,lαω1F(ω)Zn,lm,α(ω)¯dω(1-ρ2)α=alm(A)b~nl,α(B),b~nl,α(B)=1Nn,lαbnl,α(B) 92

where

alm(A)=0π02πA(n(ϑ,φ))Ylm(ϑ,φ)¯sinϑdϑdφ, 93
bnl,α(B)=01B(ρ)Rnl,α(ρ)ρ2dρ(1-ρ2)α. 94

The coefficients cn,lm,α in (92) reflect the separation of F as a product of an angular and radial factor as well as a corresponding separation of the generalized Zernike basis functions in (82). In the latter, the index l appears both in the angular and radial factor. Thus we have

A(n(ϑ,φ))=l,malmYlm(ϑ,φ), 95

while for all l=0,1,

B(ρ)=n=l,l+2,b~nl,αRnl,α(ρ). 96

For each l, the radial functions Rnl,α with n varying are a basis for functions defined on the interval [0, 1]. For separable functions, we expand the same radial function B(ρ) for each l, and it can be shown that there is a recursion formula for the radial coefficients [43].

Wavelets

We now choose appropriate radial and angular functions for our wavelets expressed in the generalized Zernike basis.

Angular Function for the Zernike Wavelets

For the angular functions we again choose orientation distribution A(n(ϑ,φ))=GsoS2(n(ϑ,φ)) for which the spherical harmonic coefficients are given by (58). After this we apply the same transformations (Funk transform and anti-symmetrization) to obtain the angular part of the wavelet.

Flat Radial Profile for All-Scale Transform

Recall the procedure of splitting of the lowest frequencies as described in Sect. 2.1.1 resulting in filters ψ0 and ψ1. In this section we design a radial function for ψ1 which is relevant for further processing. Furthermore, we already have an analytical description for ϕ0, which we set to ϕ0=Gsρ (see Eq. (37)).

The radial function of ψ0 should therefore approximate B(ρ)=1-Gsρ(ρ) on the interval [0,ϱ] and should smoothly go to zero when approaching the edges of the interval. For the moment, we set ϱ=1 and we include the scaling later. To start, we define the function

Bα,β(ρ)=(1-ρ2)αρβ, 97

see Fig. 11a for the case α=6,β=2. For this function we have the following coefficients

bn=l+2pl,α,β=β-l2p(2α+β+l+2p+3)12(β+l+1)+α+pα+p. 98

To obtain a flatter function we multiply the function Bα,β with a second-order Taylor expansion of the reciprocal function ρ(Bα,β(ρ))-1 around the function’s maximum obtained at

ρmax=12βα+12β12, 99

see Fig. 11b. The resulting function is again a sum of functions of type (97) with different values for β, so we can find the coefficients bn=l+2pl,α for the flattened function as well. For the specific case β=2 we get the following flattened function

Bα,2flat(ρ)=Bα,2(ρ)Bα,2rec(ρ)=1Bmaxρ2(1-ρ2)α1+(α+1)32α(ρ2-ρmax2)2, 100

with Bmax=Bα,2(ρmax). For this flattened function the coefficients are given by

bn=l+2pl,flat=i=02cibn=l+2pl,α,2+2i, 101

with ci the coefficients of ρ0, ρ2 and ρ4 in the second-order Taylor series of the reciprocal. These coefficients follow from (100) and are given by

Bα,2rec(ρ)=i=02ciρ2i=c0+c1ρ2+c2ρ4,withc0c1c2=1+(α+1)32αρmax4-2(α+1)32αρmax2(α+1)32α. 102

The filters from this section are summarized in the following result:

Fig. 11.

Fig. 11

Left: Function B6,2(ρ)=(1-ρ2)6ρ2. Right: Flattened function which are obtained from Bα,β(ρ) by multiplying with the second-order Taylor approximation of its reciprocal around the function maximum: B6,2flat(ρ)=1Bmax(1+7312(ρ2-17)2)B6,2(ρ)

Result 2

(Analytic 3D-wavelets in Zernike basis) Let α>0 and let A:S2R+ be a function supported mainly in a sharp convex cone around the z-axis and symmetrically around the z-axis. Then A provides our wavelet ψ^ in the Fourier domain via Eq. (65). The real part of ψ is a tube detector and the imaginary part of ψ is an edge detector, see Fig. 7. We choose radial function g(ρ)=B6,2flat(ρρN) in Eq. (18) for ψ1 and angular function A(n(ϑ,φ))=GsoS2(n(ϑ,φ)) and expand in the generalized Zernike basis:

ψ^1(ω)=n-l=2p,l,n0cn,l0Rnl,αρρNYl0(ϑ,φ). 103

The coefficients cn,l0 follow by expanding A in spherical harmonics and B6,2flat in the radial Zernike polynomials, recall Eqs. (95) and (96). This yields al0 (Eq. (58)) and bn,lα (Eq. (101)) and coefficients cn,l0:

cn,l0=Pl(0)+1-(-1)l2al0b~n,lα. 104

The spatial wavelet is given by

ψ1(x)=n-l=2p,l,n0cn,l04πilSn,lα(2πrρN)Yl0(θ,ϕ), 105

with Sn,lα given by Eq. (89) and Ylm the spherical harmonics of Eq. (57). Then we obtain rotated filters via

ψ1,n(x)=n-l=2p,l,n0m=-ll(cn)n,lm4πilSn,lα(2πrρN)Ylm(θ,ϕ),with(cn)n,lm:=(cn(β,γ))n,lm=cn,l0D0,ml(γ,β,0). 106

Since now we do have analytical expressions for the spatial filter, in contrary to the filters from Sect. 3, we sample the filters in the spatial domain using Eq. (106). The filter is a proper wavelet with fast reconstruction property (recall Definition 1).

Experiments with the Filters and the Transform Between a 3D Image and Its Orientation Score

Before considering applications of the filters in the next section, we first compare filters obtained by DFT (Sect. 3) to filters expressed in the generalized Zernike basis (Sect. 4) and inspect the quality of the reconstruction.

Comparison of Wavelets Obtained via DFT and Analytical Expressions Using the Zernike Basis

First we compare the filters obtained by sampling in the Fourier domain followed by a DFT (Sect. 3) to the filters obtained by expansion in the Zernike basis (Sect. 4). Settings were chosen such that the radial functions of both wavelets matched best and the same settings for the angular function were used. In Fig. 12 we show that the filters are very similar in shape. We see no major artifacts caused by sampling followed by an inverse DFT.

Fig. 12.

Fig. 12

Comparison of the filters obtained by sampling in the Fourier domain and performing an inverse DFT (Result 1 in Sect. 3) and the filters expressed in the generalized Zernike basis (Result 2 in Sect. 4). Left: Iso-contour plot of the filter aligned with the x-axis showing one positive iso-contour (orange) and one negative iso-contour (blue). Middle: Cross section of the filter for z=0. Right: The low-pass filter. Top Filters according to Result 1 with parameters sρ=12(1.9)2 and γ=0.85. Bottom: The filters according to Result 2 with α=3 and β=2. Both have so=12(0.4)2 and are evaluated on a grid of 31×31×31 voxels (Color figure online)

Quality of the Reconstruction

A visual inspection of the reconstruction after the transformation and reconstruction procedure can be found in Fig. 13. As expected, a small amount of regularization is observed. We see no qualitative differences between the two reconstructions.

Fig. 13.

Fig. 13

Comparison of construction and reconstruction of data A.1 using the different types of filters with the same settings as in Fig. 12. In each row, from left to right, an iso-contour of the data and 3 slices through the center of the data along the three principal axis. Top: The original data. Middle: the data after construction and reconstruction using the filters from Result 1. Bottom: the data after construction and reconstruction using the filters from Result 2

Applications

Next we present two applications of 3D image processing via invertible orientation score transforms. Figure 14 summarizes the different datasets on which we validate this kind of processing. In the invertible orientation score transform used in all experiments, we choose to use the wavelets from Result 1 and the default values of Table 1, unless stated otherwise. The reason for only using the wavelets from Result 1 is that the wavelets from Result 2 are over 10 times slower in our current implementations, partly due to the fact that the series in (106) is converging, but unfortunately not rapidly converging.

Fig. 14.

Fig. 14

Overview of the datasets used in our experiments

Table 1.

Default values for the parameters of the orientation score transform used in the application section

Parameter Default value Defining Eq.
No 42 (39)
γ 0.85 (51)
σerf 13(ρN-ϱ) (50)
sρ 12(16)2 (18)
so 12(0.45)2 (52)
Discrete wavelet size 11×11×11

Diffusion Via Invertible Orientation Scores

Background and Related Methods

Many methods exist for enhancing elongated structures based on nonlinear diffusion equations. Coherence-enhancing diffusion (CED) filtering [69] uses the structure tensor to steer the diffusion process to mainly apply diffusion along the elongated structures, therefore preserving the edges. One downside of this method is that at situations where multiple oriented structures occur at the same position, one of the structures gets destroyed. This renders this method not suitable for crossing structures and in 3D data bifurcating vessels. Interesting extensions dealing with crossings by analyzing the environment using higher-order derivatives have been proposed [63].

Methods that deal with crossings by applying coherence-enhancing diffusion via 2D orientation scores have been developed for 2D data [37, 64]. Here, we propose an extension of coherence-enhancing diffusion via 3D orientation scores to enhance elongated structures, while preserving crossings and bifurcating vessels. Preliminary results on artificial data have been shown in [32]. Here we show first results on real data, quantify the results and furthermore add additional adaptivity to the diffusion equation.

CEDOS

We now use the invertible orientation score transformation to perform data enhancement according to Fig. 2. Because R3×S2 is not a Lie group, it is common practice to embed the space of positions and orientations in the Lie group of positions and rotations SE(3) by setting

U~(x,R)=U(x,R·ez),U(x,n)=U~(x,Rn), 107

with Rn any rotation for which Rn·ez=n. The operators Φ which we consider are scale spaces on SE(3) (diffusions), and are given by Φ=Φt with

Φt(U)(y,n)=W~(y,Rn,t). 108

Here W~ is the solution of a nonlinear diffusion equation:

W~t(g,t)=i,j=16Ai|gDij(U~)Aj|gW~(g,t),W~(g,0)=Wψ1[f](x,R·ez),g=(x,R) 109

where in coherence-enhancing diffusion on orientation scores (CEDOS) Dij is adapted locally to initial condition W~(g,0) based on exponential curve fit (see [32]), and with Ai|g=(Lg)Ai|e the left-invariant vector fields on SE(3). The diffusion is better understood in locally adaptive frame {Bi}i=16. Here B3 follows from an exponential curve fits and points along our structure. B1 and B2 span the plane spatially perpendicular to our structure, and B4,B5 and B6 correspond to angular diffusion (embedding in SE(3) leads to a third angular dimension). Our diffusion then takes the diagonal form

W~t(g,t)=i=16Dii(U~)Bi|g2W~(g,t)=D11(U~)B1|g2+B2|g2W~(g,t)+D33(U~)B3|g2W~(g,t)+D44B4|g2+B5|g2+B6|g2W~(g,t) 110

where we limit ourselves to diffusion of type D11=D22, and D44=D55=D66 to preserve the data symmetry of Eq. (107). The diffusion system in Eq. (110) has been set up in previous work [45] with constant D11 and D33.

In this work we include further adaptivity by making the constants D11 and D33 data adaptive. We aim to enhance oriented structures and reduce noise as much as possible. Therefore, non-oriented regions are smoothed isotropically by setting

(D11(U~))(g)=1-exp-c1s(U~)(g)2, 111

with c1 a constant automatically set to the 50% quantile of s(U~) and s(U~)(g) a measure for orientation confidence given by minus the Laplacian in the space orthogonal to the structure orientation B3:

s(U~)(g)=-i{1,2,4,5,6}Bi|g2U~(g). 112

Since we want to stop diffusion when reaching the end of a structure, we set

(D33(U~))(g)=1-exp-c2B3|gU~(g)2, 113

with c2 a constant automatically set to the 50% quantile of |B3|g(U~)|.

We then obtain Euclidean invariant image processing via

Υf=Wψ-1,extΦWψf=Wψ-1PψΦWψf, 114

which includes inherent projection Pψ of orientation scores, even if Φ=Φt maps outside of the space of orientation scores. We write Wψ-1,ext because we extend the inverse to L2(R3×S2).

Quantification Via Contrast-to-Noise Ratio

In the next section, we quantify the denoising capabilities of the diffusion in Eq. (110) with the contrast-to-noise ratio. Next we will explain how we calculate this measure for our real medical image data.

For signal f with noise N the noisy data is given by:

fN(x)=f(x)+N(x). 115

Given such data, the noise is quantified via the contrast-to-noise ratio (CNR):

CNR(fN,f)=maxxf(x)-minxf(x)σ(f-fN), 116

where σ(f-fN) denotes the standard deviation of the difference signal f-fN, and where the numerator denotes the contrast in our data.

For real data we do not have a ground truth f but only noise signal fN and we will use the following estimation for the standard deviation of the noise and the contrast of the signal. First we estimate the contrast by determining the average value over manually segmented parts of the vessel given by region ΩS and background regions given by ΩB:

μS=fN|ΩS,μB=fN|ΩB. 117

For estimating the noise of the signal, we select regions for which the signal f can be expected to be constant (see Fig. 15). Given such a region ΩB we estimate the noise standard deviation by

σN=σ(fN|ΩB). 118

The contrast-to-noise ratio (CNR) is then given by

CNR(fN)=μS-μBσN, 119

where the numerator denotes the contrast in our data.

Fig. 15.

Fig. 15

Selected regions for determining the contrast-to-noise ratios for dataset B.1. Left: Grid containing slices of the data (top row), the same slices with segmented vessel parts (second row) and the slices with three selected background regions (third row), the slices after applying CEDOS. Right: 3D visualization of the data

Results on Cone Beam CT Data of the Abdomen

We tested our method on real Cone Beam CT data of the abdomen (Fig. 14b). The data were acquired using a Philips Allura Xper FC20 system, using a Cone Beam CT backprojection algorithm (XperCT) to generate the final volumetric image.

To quantify our method, we segmented the vessels and selected background regions, see Fig. 15. We then applied CEDOS with different end times and computed the CNR for these different end times ranging from 1 to 6, see Fig. 17. Diffusion constants D11 and D33 were determined using Eqs. (111), (113), and we set D11=0.001. For the orientation score transformation, we used so=12(0.45)2 and sρ=12(16)2. For CED we used the following settings: α=0.2 and c the 50%-quantile of κ (see [69]).

Fig. 17.

Fig. 17

Results of coherence-enhancing diffusion via orientation scores on dataset B.1, see Eq. (110). Top: CEDOS for different amounts of diffusion time. Middle: Result for CED. Bottom: Result for isotropic Gaussian regularization. For all datasets, we show one iso-contour at μBG+0.7(μFG-μBG), where μBG and μFG are the mean of the background and foreground in the (processed) data determined using the selected regions used for determining the CNR. For a better impression of the full volume, see Fig. 18. We plot results for three different times according to Fig. 16, where case (a) corresponds to optimal diffusion time for Gaussian regularization and case (c) corresponds to optimal CEDOS which is also approximately equal to optimal CED time. We see that CEDOS preserves the complex vascular geometry

As one can expect, in all cases we recognize a peak since initially noise is reduced whereas later also contrast is reduced. Compared to Gaussian diffusion and CED, we reach a higher CNR (Fig. 16). Furthermore, in the CEDOS and CED case, the CNR does not decrease much when applying more diffusion making it more robust with respect to choice in diffusion time. The fact that CED does not achieve high CNR ratios and performs relatively bad in this test is that the diffusion matrix is not designed to reduce background noise (which is used here to quantify noise) but mainly to enhance orientated patterns; for this reason we also set α relatively high to still achieve noise reduction in the background regions.

Fig. 16.

Fig. 16

Contrast-to-noise ratio (CNR) against diffusion time for CEDOS compared to Gaussian regularization and CED of data B.1 depicted in Fig. 15. The times denoted by a, b and c correspond to the diffusion times shown in Fig. 17

For 3D visualization of the diffusion results for optimal diffusion time (determined from the CNR), see Fig. 18. Here we see that compared to Gaussian diffusion our anisotropic diffusion reduced more noise while still maintaining the important structures. A similar thing is achieved by anisotropic diffusion in CED but bifurcating vessels are destroyed by this method, as in this method the diffusion is mainly performed along the orientation of one of the vessels at the bifurcation (see the black circles in Fig. 18).

Fig. 18.

Fig. 18

Volume rendering of the diffusion results for datasets B.1, B.2 and B.3 (from top to bottom) visualized in the Philips viewer [62] using default settings in all cases. For all cases we used optimized diffusion time according to Fig. 16

Influence of CEDOS on Vessel Edge Location

In order to test the influence of our regularization method on vessel features, we implemented a simple edge detection algorithm. We manually selected positions in the data and detect the edges in the vessel cross sections by extracting radial profiles from the centerline outward and looking for the minimum in first-order gaussian derivative. Compared to Gaussian regularization, the vessel edge position is not influenced by our regularization method (Fig. 19), which is highly important for applications which rely on accurate vessel lumen measurements, e.g., stent positioning and navigation of endovascular devices. The key explanation for this benefit is that at the vessel we get a very low D11 (Eq. 111) and therefore we smooth only along the vessel.

Fig. 19.

Fig. 19

Measurement of vessel radius in vessel cross sections after different amounts of diffusion in dataset B.1. Top left: 3D Visualization of the data with the selected slices. Top middle: Radii measurements for increasing diffusion time for CEDOS. Top right: Radii measurements for increasing diffusion time for Gaussian regularization. The detected vessel width is not influenced by our regularization method while this does occur for Gaussian regularization. Bottom: Cross sections of one vessel for increasing diffusion time with detected vessel edge positions (green points) and search area for edge detection (red circle) (Color figure online)

Tubularity Measure

Background and Related Methods

In this section we propose a tubularity measure based on the edge information in our orientation scores. This tubularity measure is then used for vessel width measurements and could be used for vessel segmentation.

Tubularity measures are designed to have a high response on the centerline of tubular structures. One such tubularity measure is the image gradient flux filter [14] which is used for vessel segmentation. An extension of the gradient flux filter is the optimally oriented flux filter [47] which introduces the notion of oriented flux making the filter orientation sensitive.

A 2D tubularity measure based on orientation scores was proposed in [9]. The advantage of this tubularity measure is that it included nonlinearity and that the implementation via orientation scores still has a response at crossing vessels. Here we propose an extension of this tubularity measure to 3D making use of 3D orientation scores.

Tubularity Via Orientation Scores

For the tubularity measure we detect edges in the plane perpendicular to orientation n. Within this plane, the product of two opposite edges at radius r and in-plane orientation θ at position xR3 is given by

Eprod(x,n,r,θ)=Im+[U(x+rn(θ),n(θ))]·Im+[U(x-rn(θ),-n(θ))], 120

where Im+(z)=max{0,Im(z)} and n(θ)=cosθe1+sinθe2, with {e1,e2} an orthogonal basis for the orthogonal complement of n=span{n}. The product of the two Im+ edge responses in Eq. (120) yields a better performance than taking the sum, as is done in [9, Fig. 12.2] and [19]. The idea behind taking the product instead of the sum is that we need a high edge response in both directions. See Fig. 20 for a schematic visualization. Since for a real tube all edges should be present, and we do no want any response for, e.g., plate structures, we take a minimum over the perpendicular orientations parametrized by θ. This is done as follows

V(x,n,r)=minθ[0,2π)0πKσoVor(θ-θ)Eprod(x,n,r,θ)dθ. 121

For the angular regularization kernel we use the simple 2π periodic 1D-diffusion kernel

Fig. 20.

Fig. 20

Schematic visualization of the edges used in the tubularity measure V(x,n,r). Left: A 3D iso-contour visualization of a vessel with orientation n. The coordinates (θ,r) are polar coordinates for the plane perpendicular to orientation n spanned by e1 and e2. Right: In this plane opposite edges are multiplied in Eprod(x,n,r,θ). The edge is expected to have outward orientation n(θ)

KσoVor(θ)=kZGS1σoV(θ+2kπ)=kZ12πσove-|θ+2kπ|22|σoV|2 over θ, instead of the true diffusion kernel (72), where for practical values we use σoV=π/8 the series of rapidly converging kernels can be reasonably truncated already at the first term k=0.

From the tubularity measure (121), we extract the following features:

st(x)=maxnS2,rR+V(x,n,r), 122
n(x)=argmaxnS2maxrR+V(x,n,r), 123
r(x)=argmaxrR+maxnS2V(x,n,r). 124

Here st(x) is the tubularity confidence which is a measure for how certain we are at least one tubular structure is present at position x. The features n(x) and r(x) are the orientation and radius of optimal tubularity response at position x.

Results on Artificial Data

For the validation of our tubularity measure, we constructed 18 artificial datasets with a random tubular structure with randomly varying radius (Fig. 14c). For the tubularity measure we used the following settings: σoV=π/8 and we discretized the θ-integral in Eq. (121) using 8 orientations.

As validation we compared the optimal radius to the ground truth radius and inspect the tubularity confidence. The transversal profiles in the tubes are modeled by Gaussians, and we define the ground truth radius rGT(x)=σ(x) at the standard deviation σ(x) of the Gaussian. Thereby, the tube boundaries are at the inflection point of the transversal Gaussian profiles.

The tubularity confidence is selective on the vessel centerline, and the found optimal radius is a reasonable estimation of the ground truth radius, see Figs. 21 and 22, and follows the correct trend although outliers typically tend to produce an overestimate.

Fig. 21.

Fig. 21

Tubularity on artificial datasets and comparison of measured radius against ground truth radius for datasets C.1, C.2 and C.3 (top to bottom). Left: The data. Middle left: The centerline with ground truth radius in color. Middle right: The tubularity confidence st(x) (max of tubularity over radius and orientation) with radius of max response r(x) in color. Right: Measured radius r(x) against ground truth radius rGT(x) on the ground truth centerline. The opacity of the plotted points is linearly scaled with the tubularity confidence

Fig. 22.

Fig. 22

Measured radius r(x) against ground truth radius on the ground truth centerline for all 18 datasets. The opacity of the plotted points is linearly scaled with the tubularity confidence

Results on 3D Rotational Angiography Data of the Brain in Patients with Arteriovenous Malformation (AVM)

We applied the tubularity measure to 3D rotational angiography of the brain in patients with arteriovenous malformation (Fig. 14a). The data were acquired using a Philips Allura Xper FC20 system, using a 3D rotational angiography backprojection algorithm (3DRA) to generate the final volumetric image. For the tubularity measure we used the following settings: σoV=π/8 and we discretized the θ-integral in Eq. (121) using 12 orientations. For the orientation score transformation we used so=12(0.4)2,sρ=12(5)2 and evaluated on a grid of 21×21×21 pixels and default settings for the other parameters.

In Fig. 23 we show our tubularity measure for this medical data. The tubularity measure gives sharp responses for vessel centerlines. We also show optimal orientation n(x) and a simple segmentation given by the 0-iso-contour of the distance map d(x,Bci,r(ci))=mincix-ci-r(ci), where ci are the positions given by the 1% quantile of highest responses.

Fig. 23.

Fig. 23

Tubularity on real data. Top: Dataset A.1. Bottom: Dataset A.2. Left: The data. Middle left: The projected tubularity (max over radius and orientation) colored by radius. Middle right: Orientation of max response for 1% quantile of highest responses in the tubularity confidence. Right: Segmentation based on radius of max response for 1% quantile of highest responses in the tubularity confidence. The plotted surface is the 0-iso-contour of the distance map d(x,Bci,r(ci))=mincix-ci-r(ci), where ci are the positions given by the 1% quantile of highest responses

Conclusion

We presented theory and filters for the 3D orientation score transformation which is valuable in handling complex oriented structures in 3D image data. Then we showed applications of this transformation.

First, we proposed filters for a 3D orientation score transform. We presented two types of filters: The first uses a discrete Fourier transform and the second is designed in the 3D generalized Zernike basis which allowed us to find analytical expressions for the spatial filters. Both filters allowed for an invertible transformation. The filters and the quality of the reconstruction are assessed in Sect. 5, where we showed that the discrete filters approximate their analytical counterparts well. We also verified the invertibility of our transformation by showing data reconstructions of real medical data.

The orientation score transform was then used in two different applications. In the first we presented an extension of coherence-enhancing diffusion via 3D orientation scores which we applied to real 3D medical data and showed our method effectively reduced noise while maintaining the complex vessel geometry. In the second application we propose a new nonlinear tubularity measure via 3D orientation scores. The tubularity measure has sharp responses for vessel centerlines, and we showed its use in radius extraction and complex vessel segmentation.

In this work basic applications of the tubularity measure are shown. Future work would include using the tubularity measure in more advanced vessel segmentation procedures. Furthermore, many other applications exist for the 3D orientation scores and many techniques developed for 2D orientation scores can now be extended to 3D. First steps are presented in this paper, where the extension of 2D CEDOS and a 2D tubularity measure via 2D orientation scores are given. It is also interesting to explore the nonlinear diffusion procedure (Eq. (110)) for contextual processing of diffusion-weighted MRI and to compare with existing approaches [58, 68].

Acknowledgements

We would like to thank professors Jacques Moret and Laurent Spelle from the NEURI center, Department of NRI, Bicêtre University Hospital Paris, France, for providing the AVM dataset used in this article. The research leading to the results of this article has received funding from the European Research Council (ERC) under the European Community 7th Framework Programme (FP7/20072014)/ERC Grant Agreement No. 335555 (Lie Analysis).

Biographies

M. H. J. Janssen

received his M.Sc. degree (cum laude) in Biomedical Engineering in 2014 at TU/e, The Netherlands. Afterward, he worked as a Ph.D. student in the Department of Applied Mathematics and Computer Science at TU/e, working on the topic of 3D multi-orientation analysis for medical image processing. His research interests include geometrical methods for medical image analysis, multi-orientation analysis and differential equations on Lie groups. Currently, he is working for VitalHealth software, developing e-Health software applications.graphic file with name 10851_2018_806_Figa_HTML.jpg

A. J. E. M. Janssen

(1953) received his Engineering and Ph.D. degree in Mathematics from the Eindhoven University of Technology, The Netherlands, in 1976 and 1979, respectively. From 1979 to 1981 he was a Bateman Research Instructor at the Mathematics Department of California Institute of Technology. From 1981 until 2010 he worked with Philips Research Laboratories, Eindhoven, where he was a Research Fellow since 1999 and recipient of the Gilles Holst award in 2003. His principal responsibility at Philips Research was to provide high-level mathematical service and consultancy in mathematical analysis. He continues his research and consultancy activities from his current affiliations TU/e, EURANDOM, with further contacts with TU-Delft (Optics) and ASML. In 2003, he was elected Fellow of the IEEE for his contributions to mathematical time–frequency analysis and signal processing. Dr. Janssen has published over 200 papers in the fields of Fourier analysis, signal analysis, time–frequency analysis, electron microscopy, optical diffraction theory, acoustics and queueing theory.graphic file with name 10851_2018_806_Figb_HTML.jpg

E. J. Bekkers

received his M.Sc. degree in Biomedical Engineering in 2012 at Eindhoven University of Technology (TU/e), the Netherlands. In January 2017 he received his Ph.D. degree (cum laude) for his thesis "Retinal Image Analysis using Sub-Riemannian Geometry in SE(2)," which was conducted in a combined position at Biomedical Engineering at TU/e and at the medical device company i-Optics, the Hague, the Netherlands. He currently is a postdoc researcher at the Department of Mathematics and Computer Science at TU/e, working on the application of Lie group (and control) theory in medical image analysis. His research interests include differential geometry, PDEs, variational methods, machine learning, medical imaging and statistical biomarker analysis.graphic file with name 10851_2018_806_Figc_HTML.jpg

J. Oliván Bescós

(1976) received his Physics degree from University of Zaragoza, Spain, in 1999. In 2004 he finished his Ph.D. in Electrical Engineering at the University of Twente, The Netherlands, where he conducted research on medical visualization and image processing. Since 2004 he works at Philips Healthcare, where he participates in the design, development and validation of algorithms and medical software to support doctors in imaging-guided interventions. He has been part of several multi-disciplinary projects acting as a liaison between clinicians, image processing and visualization scientists and engineering teams.graphic file with name 10851_2018_806_Figd_HTML.jpg

R. Duits

received his M.Sc. degree (cum laude) in Mathematics in 2001 at the TU/e, The Netherlands. He received his Ph.D. degree (cum laude) at the Department of Biomedical Engineering at the TU/e. Now he is associate professor at the Department of Applied Mathematics & Computer Science and affiliated part time at the Biomedical Image Analysis group at the Biomedical Engineering Department. He has received an ERC-StG 2013 personal grant (Lie analysis, No. 335555). His research interests subtend group theory, functional analysis, differential geometry, PDE’s, harmonic analysis, geometric control and their applications to biomedical imaging and vision.graphic file with name 10851_2018_806_Fige_HTML.jpg

A Steerable Orientation Score Transform

In this article we often rely on a spherical harmonic decomposition of the angular part of proper wavelets in the spatial and the Fourier domain. As the choice of radial basis varies in Sect. 4, and since it does not affect steerable filter [36, 38, 61, 67] properties, we simply write {gnl(r)}n=0 for a radial orthonormal basis for L2(R+,r2dr) associated to each 2l+1-dimensional SO(3)-irreducible subspace indexed by l. Then by the celebrated Bochner–Hecke theorem [34], this induces a corresponding orthonormal radial basis {g~nl(ρ)}n=0 in the Fourier domain which can be obtained by a Hankel-type of transform [25, Ch. 7]. We expand our wavelets in spherical harmonics Ylm and ball coordinates (cf. (78)) accordingly:

ψ(x)=n=0αlngnl(r)Yl0(θ,ϕ),withx=(rcosϕsinθ,rcosϕsinθ,rcosθ),ψ^(ω)=n=0αlng~nl(ρ)Yl0(ϑ,φ),withω=(ρcosφsinϑ,ρcosφsinϑ,ρcosϑ). 125

Remark 4

In Sect. 4, we consider the modified Zernike basis in which case gnl and g~nl are given by, respectively, (85) and (82), whereas for the harmonic oscillator basis one has gnl=il(-1)n+lg~nl given, respectively, by (77) and (79).

We obtain steerability via finite series truncation at n=N and l=L. Then we rotate the steerable kernels via the Wigner-D functions D0,ml(γ,β,0)R (cf. Sect. 3.2.3) and one obtains the following steerable implementations of orientation scores:

U(x,n)=n=0Nl=0Lm=-llαlnD0,ml(γ,β,0)¯·((gnlYl0)f)(x)=n=0Nl=0Lm=-llαlnD0,ml(γ,β,0)¯·F-1g~nlYl0¯·f^(x) 126

where n=(cosγsinβ,sinγsinβ,cosβ)T, denotes correlation, the overline denotes complex conjugation and with function product (g~nlYlm)(x)=g~nl(r)Ylm(θ,ϕ).

B Table of Notations

Symbol Explanation Reference
B.1 Spaces and Input Data
R3×S2 Space of positions and orientations Page 1
L2ϱ(R3)={fL2(R3)|supp(Ff)Bϱ}, with ϱ>0 Space of frequency ball-limited images (2)
Bρ={xR3|x<ρ}, with ρ>0 The ball around the origin with radius ρ (2)
B.2 Orientation Score Transformation
Wψ Orientation score transformation (3)
Wψ-1 Data reconstruction via the inverse orientation score transformation (5)
ψ Wavelet used for the orientation score transformation (3)
ψ^ Fourier transform of the wavelet used for the orientation score transformation (6)
Mψ Factor used to quantify the stability of the transformation (6)
Nψ Factor used to quantify the stability of the transformation when using the simplified reconstruction by integration (12)
Wψd,(Wψd)-1,Mψd,Nψd Discretized versions of Wψ,(Wψ)-1,Mψ,Nψ (39), (40), (41), (42)
dσ and Δi Spherical area measure and discretized spherical area measure (5), (40)
B.3 Coordinates
x=(x,y,z) Cartesian coordinates real space -
ω=(ωx,ωy,ωz) Cartesian coordinates Fourier domain -
(r,θ,ϕ),x=(rsinθcosϕ,rsinθsinϕ,rcosθ) Spherical coordinates real space (49)
(ρ,ϑ,φ),ω=(ρsinϑcosφ,ρsinϑsinφ,ρcosϑ) Spherical coordinates Fourier domain (49)
B.4 Wavelets
g Radial function of the cake filters (50)
A Orientation distribution used in wavelet construction (52)
Ylm Spherical harmonics (57)
Zn,lm,α 3D generalized Zernike functions (82)
Rnl,α Radial part of the 3D generalized Zernike functions (84)
Sn,lα Radial part of the inverse Fourier transform 3D generalized Zernike functions (89)
B.5 Applications
V(x,n,r) Tubularity measure (121)
st(x) Tubularity confidence (122)
n(x) Orientation of maximum tubularity response (122)
r(x) Radius of maximum tubularity response (122)

Footnotes

1

In practice we choose the default values δ=18 and M=1.1 and ε=0.01 and note it is actually the ratio Mδ that determines the condition number. It is just a convenient choice to set the upper bound close to 1.

2

For our applications this numerical approach in [18] was sufficient. There are alternative approaches [40] with profound analysis on optimized spherical samplings minimizing squared quadrature errors in Fourier transforms.

3

This can be observed in Fig. 6, where we recognize a planar derivative filter in the direction of n because it is anti-symmetric and has a mainly positive and mainly negative side.

4

The series in (81) converges point-wise but not in L2-sense. So for taking the limit L one must rely on a distributional wavelet transform, since ψHLL2(R3) but ψHLH-4(R3), i.e., it is contained in the dual of the 4th-order Sobolev space. Such technicalities do not occur in the Zernike basis as we will see later in Sect. 4.

Contributor Information

M. H. J. Janssen, Phone: +31-40-2478897, Email: M.H.J.Janssen@tue.nl

A. J. E. M. Janssen, Email: A.J.E.M.Janssen@tue.nl

E. J. Bekkers, Email: E.J.Bekkers@tue.nl

J. Oliván Bescós, Email: Javier.olivan.bescos@philips.com.

R. Duits, Phone: +31-40-2472859, Email: R.Duits@tue.nl

References

  • 1.Aarts RM, Janssen AJEM. On-axis and far-field sound radiation from resilient flat and dome-shaped radiators. J. Acoust. Soc. Am. 2009;125(3):1444–1455. doi: 10.1121/1.3075594. [DOI] [PubMed] [Google Scholar]
  • 2.Ali ST. A general theorem on square-integrability: vector coherent states. J. Math. Phys. 1998;39(8):3954. doi: 10.1063/1.532478. [DOI] [Google Scholar]
  • 3.Ali ST, Antoine JP, Gazeau JP. Coherent States, Wavelets, and Their Generalizations. Graduate Texts in Contemporary Physics. New York: Springer; 2000. [Google Scholar]
  • 4.Antoine JP, Murenzi R, Vandergheynst P. Directional wavelets revisited: Cauchy wavelets and symmetry detection in patterns. Appl. Comput. Harmon. Anal. 1999;6:314–345. doi: 10.1006/acha.1998.0255. [DOI] [Google Scholar]
  • 5.Barbieri D, Citti G. Reproducing kernel Hilbert spaces of CR functions for the Euclidean motion group. Anal. Appl. 2014;13(03):331–346. doi: 10.1142/S021953051450047X. [DOI] [Google Scholar]
  • 6.Barbieri D, Citti G, Cocci G, Sarti A. A cortical-inspired geometry for contour perception and motion integration. J. Math. Imaging Vis. 2014;49(3):511–529. doi: 10.1007/s10851-013-0482-z. [DOI] [Google Scholar]
  • 7.Batard T, Sochen N. A class of generalized laplacians on vector bundles devoted to multi-channel image processing. J. Math. Imaging Vis. 2014;48(3):517–543. doi: 10.1007/s10851-013-0426-7. [DOI] [Google Scholar]
  • 8.Bekkers E, Duits R, Mashtakov A, Sanguinetti G. A PDE approach to data-driven sub-Riemannian geodesics in SE(2) SIAM J. Imaging Sci. 2015;8(4):2740–2770. doi: 10.1137/15M1018460. [DOI] [Google Scholar]
  • 9.Bekkers, E.J.: Retinal Image Analysis Using Sub-Riemannian Geometry in SE(2). Ph.D. Thesis, University of Technology, Eindhoven (2017)
  • 10.Bekkers EJ, Duits R, Berendschot T, ter Haar Romeny BM. A multi-orientation analysis approach to retinal vessel tracking. J. Math. Imaging Vis. 2014;49(3):583–610. doi: 10.1007/s10851-013-0488-6. [DOI] [Google Scholar]
  • 11.Bekkers, E.J., Zhang, J., Duits, R., ter Haar Romeny, B.M.: Curvature based biomarkers for diabetic retinopathy via exponential curve fits in SE(2). In: Proceedings of the Ophthalmic Medical Image Analysis First International Workshop, pp. 113–120 (2015)
  • 12.Born M, Wolf E. Principles of Optics. 7. Cambridge: Cambridge University Press; 1999. [Google Scholar]
  • 13.Boscain U, Chertovskih RA, Gauthier JP, Remizov AO. Hypoelliptic diffusion and human vision: a semidiscrete new twist. SIAM J. Imaging Sci. 2014;7(2):669–695. doi: 10.1137/130924731. [DOI] [Google Scholar]
  • 14.Bouix S, Siddiqi K, Tannenbaum A. Flux driven automatic centerline extraction. Med. Image Anal. 2005;9(3):209–221. doi: 10.1016/j.media.2004.06.026. [DOI] [PubMed] [Google Scholar]
  • 15.Brunner TA. Impact of lens aberrations on optical lithography. IBM J. Res. Dev. 1997;41(1.2):57–67. doi: 10.1147/rd.411.0057. [DOI] [Google Scholar]
  • 16.Burgeth B, Didas S, Weickert J. A general structure tensor concept and coherence-enhancing diffusion filtering for matrix fields. In: Laidlaw D, Weickert J, editors. Visualization and Processing of Tensor Fields, Mathematics and Visualization. Berlin: Springer; 2009. pp. 305–323. [Google Scholar]
  • 17.Burgeth B, Pizarro L, Didas S, Weickert J. 3D-coherence-enhancing diffusion filtering for matrix fields. In: Florack L, Duits R, Jongbloed G, van Lieshout MC, Davies L, editors. Mathematical Methods for Signal and Image Analysis and Representation, Computational Imaging and Vision. London: Springer; 2012. pp. 49–63. [Google Scholar]
  • 18.Caruyer E, Lenglet C, Sapiro G, Deriche R. Design of multishell sampling schemes with uniform coverage in diffusion MRI. Magn. Reson. Med. 2013;69(6):1534–1540. doi: 10.1002/mrm.24736. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Chen, D., Cohen, L.D.: Automatic vessel tree structure extraction by growing minimal paths and a mask. In: 2014 IEEE 11th International Symposium on Biomedical Imaging (ISBI), pp. 802–805 (2014)
  • 20.Chung, M.K.: Heat kernel smoothing on unit sphere. In: 3rd IEEE International Symposium on Biomedical Imaging: Nano to Macro, 2006, pp. 992–995 (2006)
  • 21.Citti G, Franceschiello B, Sanguinetti G, Sarti A. Sub-Riemannian mean curvature flow for image processing. SIAM J. Imaging Sci. 2016;9(1):212–237. doi: 10.1137/15M1013572. [DOI] [Google Scholar]
  • 22.Citti G, Sarti A. A cortical based model of perceptual completion in the roto-translation space. J. Math. Imaging Vis. 2006;24(3):307–326. doi: 10.1007/s10851-005-3630-2. [DOI] [Google Scholar]
  • 23.Descoteaux M, Angelino E, Fitzgibbons S, Deriche R. Regularized, fast, and robust analytical Q-ball imaging. Magn. Reson. Med. 2007;58(3):497–510. doi: 10.1002/mrm.21277. [DOI] [PubMed] [Google Scholar]
  • 24.Dirksen P, Braat JJM, Janssen AJEM, Juffermans CAH. Aberration retrieval using the extended Nijboer–Zernike approach. J. Microlithogr. Microfabr. Microsyst. 2003;2:61–67. [Google Scholar]
  • 25.Duits, R.: Perceptual Organization in Image Analysis. Ph.D. Thesis, Technische Universiteit Eindhoven (2005)
  • 26.Duits, R., Burgeth, B.: Scale spaces on lie groups. In: Scale Space and Variational Methods in Computer Vision, pp. 300–312. Springer, Berlin (2007)
  • 27.Duits R, Duits M, van Almsick M, ter Haar Romeny BM. Invertible orientation scores as an application of generalized wavelet theory. PRIA. 2007;17(1):42–75. [Google Scholar]
  • 28.Duits R, Felsberg M, Granlund G, ter Haar Romeny BM. Image analysis and reconstruction using a wavelet transform constructed from a reducible representation of the euclidean motion group. Int. J. Comput. Vis. 2007;72(1):79–102. doi: 10.1007/s11263-006-8894-5. [DOI] [Google Scholar]
  • 29.Duits R, Franken E. Left-invariant diffusions on the space of positions and orientations and their application to crossing-preserving smoothing of HARDI images. Int. J. Comput. Vis. 2010;92(3):231–264. doi: 10.1007/s11263-010-0332-z. [DOI] [Google Scholar]
  • 30.Duits R, Franken E. Left-invariant parabolic evolutions on SE(2) and contour enhancement via invertible orientation scores Part I: linear left-invariant diffusion equations on SE(2) Q. Appl. Math. 2010;68(2):255–292. doi: 10.1090/S0033-569X-10-01172-0. [DOI] [Google Scholar]
  • 31.Duits R, Führ H, Janssen B, Bruurmijn M, Florack L, van Assen H. Evolution equations on Gabor transforms and their applications. Appl. Comput. Harmon. Anal. 2013;35(3):483–526. doi: 10.1016/j.acha.2012.11.007. [DOI] [Google Scholar]
  • 32.Duits R, Janssen MHJ, Hannink J, Sanguinetti GR. Locally adaptive frames in the roto-translation group and their applications in medical imaging. J. Math. Imaging Vis. 2016;56(3):367–402. doi: 10.1007/s10851-016-0641-0. [DOI] [Google Scholar]
  • 33.Dunkl CF. Orthogonal Polynomials of Several Variables. Cambridge: Cambridge University Press; 2014. [Google Scholar]
  • 34.Faraut J, Harzallah K. Deux Cours d’Analyse Harmonique. Basel: Birkhauser; 1987. [Google Scholar]
  • 35.Felsberg Michael. Computational Imaging and Vision. London: Springer London; 2012. Adaptive Filtering Using Channel Representations; pp. 31–48. [Google Scholar]
  • 36.Franken, E.: Enhancement of Crossing Elongated Structures in Images. Ph.D. Thesis, Technische Universiteit Eindhoven (2008)
  • 37.Franken E, Duits R. Crossing-preserving coherence-enhancing diffusion on invertible orientation scores. Int. J. Comput. Vis. 2009;85(3):253–278. doi: 10.1007/s11263-009-0213-5. [DOI] [Google Scholar]
  • 38.Freeman WT, Adelson EH. The design and use of steerable filters. IEEE Trans. Pattern Anal. Mach. Intell. 1991;13(9):891–906. doi: 10.1109/34.93808. [DOI] [Google Scholar]
  • 39.Führ Hartmut. Lecture Notes in Mathematics. Berlin, Heidelberg: Springer Berlin Heidelberg; 2005. 2. Wavelet Transforms and Group Representations; pp. 15–58. [Google Scholar]
  • 40.Gräf M, Potts D. On the computation of spherical designs by a new optimization approach based on fast spherical fourier transforms. Numer. Math. 2011;119(4):699–724. doi: 10.1007/s00211-011-0399-7. [DOI] [Google Scholar]
  • 41.Griffiths DJ. Introduction to Quantum Mechanics. 2. Cambridge: Cambridge University Press; 2016. [Google Scholar]
  • 42.Isham CJ, Klauder JR. Coherent states for n-dimensional Euclidean groups E(n) and their application. J. Math. Phys. 1991;32(3):607–620. doi: 10.1063/1.529402. [DOI] [Google Scholar]
  • 43.Janssen, A.J.: Generalized 3d Zernike Functions for Analytic Construction of Band-Limited Line-Detecting Wavelets. ArXiv preprint arXiv:1510.04837 (2015)
  • 44.Janssen, M.H.J., Duits, R., Breeuwer, M.: Invertible orientation scores of 3D images. In: Scale Space and Variational Methods in Computer Vision, Lecture Notes in Computer Science, vol. 9087, pp. 563–575. Springer (2015)
  • 45.Janssen Michiel H. J., Dela Haije Tom C. J., Martin Frank C., Bekkers Erik J., Duits Remco. Lecture Notes in Computer Science. Cham: Springer International Publishing; 2017. The Hessian of Axially Symmetric Functions on SE(3) and Application in 3D Image Analysis; pp. 643–655. [Google Scholar]
  • 46.Kalitzin SN, ter Haar Romeny BM, Viergever MA. Invertible apertured orientation filters in image analysis. Int. J. Comput. Vis. 1999;31(2–3):145–158. doi: 10.1023/A:1008013815039. [DOI] [Google Scholar]
  • 47.Law, M.W.K., Chung, A.C.S.: Three dimensional curvilinear structure detection using optimally oriented flux. In: Computer Vision—ECCV 2008, pp. 368–382. Springer, Berlin (2008)
  • 48.Lee TS. Image representation using 2D Gabor wavelets. IEEE Trans. Pattern Anal. Mach. Intell. 1996;18(10):959–971. doi: 10.1109/34.541406. [DOI] [Google Scholar]
  • 49.Louis A. Orthogonal function series expansions and the null space of the Radon transform. SIAM J. Math. Anal. 1984;15(3):621–633. doi: 10.1137/0515047. [DOI] [Google Scholar]
  • 50.Louis A, Maass D, Rieder A. Wavelets: Theory and Application. New York: Wiley; 1997. [Google Scholar]
  • 51.Mallat S. A Wavelet Tour of Signal Processing. London: Academic Press; 1999. [Google Scholar]
  • 52.Mashtakov A, Duits R, Sachkov Y, Bekkers EJ, Beschastnyi I. Tracking of lines in spherical images via sub-Riemannian geodesics in SO(3) J. Math. Imaging Vis. 2017;58(2):239–264. doi: 10.1007/s10851-017-0705-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 53.Mashtakov AP, Ardentov AA, Sachkov YL. Parallel algorithm and software for image inpainting via sub-Riemannian minimizers on the group of rototranslations. Numer. Math. Theory Methods Appl. 2013;6(01):95–115. [Google Scholar]
  • 54.Muhlich M, Aach T. Analysis of multiple orientations. IEEE Trans. Image Process. 2009;18(7):1424–1437. doi: 10.1109/TIP.2009.2019307. [DOI] [PubMed] [Google Scholar]
  • 55.Olver FWJ. NIST Handbook of Mathematical Functions. Cambridge: Cambridge University Press; 2010. [Google Scholar]
  • 56.Pechaud, M., Keriven, R., Peyre, G.: Extraction of tubular structures over an orientation domain. In: 2009 IEEE Conference on Computer Vision and Pattern Recognition, pp. 336–342 (2009)
  • 57.Petitot J. The neurogeometry of pinwheels as a sub-Riemannian contact structure. J. Physiol. 2003;97(2–3):265–309. doi: 10.1016/j.jphysparis.2003.10.010. [DOI] [PubMed] [Google Scholar]
  • 58.Portegies JM, Fick RHJ, Sanguinetti GR, Meesters SPL, Girard G, Duits R. Improving fiber alignment in hardi by combining contextual pde flow with constrained spherical deconvolution. PLoS ONE. 2015;10(10):1–33. doi: 10.1371/journal.pone.0138122. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 59.Prandi Dario, Boscain Ugo, Gauthier Jean-Paul. Lecture Notes in Computer Science. Cham: Springer International Publishing; 2015. Image Processing in the Semidiscrete Group of Rototranslations; pp. 627–634. [Google Scholar]
  • 60.Rajan, S., Das, T., Krishnakumar, R.: An analytical method for the detection of exudates in retinal images using invertible orientation scores. In: Proceedings of the World Congress on Engineering, vol. 1 (2016)
  • 61.Reisert, M.: Group Integration Techniques in Pattern Analysis—A Kernel View. Ph.D. Thesis, Albert-Ludwigs- University (2008)
  • 62.Ruijters D, Vilanova A. Optimizing GPU volume rendering. J. WSCG. 2006;14(1–3):9–16. [Google Scholar]
  • 63.Scharr Hanno. Lecture Notes in Computer Science. Berlin, Heidelberg: Springer Berlin Heidelberg; 2006. Diffusion-Like Reconstruction Schemes from Linear Data Models; pp. 51–60. [Google Scholar]
  • 64.Sharma U, Duits R. Left-invariant evolutions of wavelet transforms on the similitude group. Appl. Comput. Harmon. Anal. 2015;39(1):110–137. doi: 10.1016/j.acha.2014.09.001. [DOI] [Google Scholar]
  • 65.Sifre, L., Mallat, P.S.: Rigid-Motion Scattering For Image Classification. Ph.D. Thesis, Ecole Polytechnique, CMAP, Paris (2014)
  • 66.Steidl Gabriele, Teuber Tanja. Lecture Notes in Computer Science. Berlin, Heidelberg: Springer Berlin Heidelberg; 2009. Anisotropic Smoothing Using Double Orientations; pp. 477–489. [Google Scholar]
  • 67.van Almsick, M.: Context Models of Lines and Contours. Ph.D. Thesis, Technische Universiteit Eindhoven (2007)
  • 68.Vogt, T., Lellmann, J.: Measure-Valued Variational Models with Applications to Diffusion-Weighted Imaging. ArXiv preprint arXiv:1710.00798 (2017)
  • 69.Weickert J. Coherence-enhancing diffusion filtering. Int. J. Comput. Vis. 1999;31(2–3):111–127. doi: 10.1023/A:1008009714131. [DOI] [Google Scholar]
  • 70.Zhang J, Dashtbozorg B, Bekkers E, Pluim JPW, Duits R, ter Haar Romeny BM. Robust retinal vessel segmentation via locally adaptive derivative frames in orientation scores. IEEE Trans. Med. Imaging. 2016;35(12):2631–2644. doi: 10.1109/TMI.2016.2587062. [DOI] [PubMed] [Google Scholar]
  • 71.Zweck J, Williams LR. Euclidean group invariant computation of stochastic completion fields using shiftable-twistable functions. J. Math. Imaging Vis. 2004;21(2):135–154. doi: 10.1023/B:JMIV.0000035179.47895.bc. [DOI] [Google Scholar]

Articles from Journal of Mathematical Imaging and Vision are provided here courtesy of Springer

RESOURCES