Skip to main content
PLOS One logoLink to PLOS One
. 2016 May 12;11(5):e0154160. doi: 10.1371/journal.pone.0154160

Filter Design and Performance Evaluation for Fingerprint Image Segmentation

Duy Hoang Thai 1,2,*, Stephan Huckemann 1, Carsten Gottschlich 1
Editor: M Sohel Rahman3
PMCID: PMC4865205  PMID: 27171150

Abstract

Fingerprint recognition plays an important role in many commercial applications and is used by millions of people every day, e.g. for unlocking mobile phones. Fingerprint image segmentation is typically the first processing step of most fingerprint algorithms and it divides an image into foreground, the region of interest, and background. Two types of error can occur during this step which both have a negative impact on the recognition performance: ‘true’ foreground can be labeled as background and features like minutiae can be lost, or conversely ‘true’ background can be misclassified as foreground and spurious features can be introduced. The contribution of this paper is threefold: firstly, we propose a novel factorized directional bandpass (FDB) segmentation method for texture extraction based on the directional Hilbert transform of a Butterworth bandpass (DHBB) filter interwoven with soft-thresholding. Secondly, we provide a manually marked ground truth segmentation for 10560 images as an evaluation benchmark. Thirdly, we conduct a systematic performance comparison between the FDB method and four of the most often cited fingerprint segmentation algorithms showing that the FDB segmentation method clearly outperforms these four widely used methods. The benchmark and the implementation of the FDB method are made publicly available.

Introduction

Nowadays, fingerprint recognition is used by millions of people in their daily life for verifying a claimed identity in commercial applications ranging from check-in at work places or libraries, access control at amusement parks or zoos to unlocking notebooks, tablets or mobile phones. Most fingerprint recognition systems are based on minutiae as features for comparing fingerprints [1]. Typical processing steps prior to minutiae extraction are fingerprint segmentation, orientation field estimation and image enhancement. The segmentation step divides an image into foreground, the region of interest (ROI), and background. Two types of error can occur in this step and both have a negative impact on the recognition rate: ‘true’ foreground can be labelled as background and features like minutiae can be lost, or ‘true’ background can be misclassified as foreground and spurious features may be introduced. It is desirable to have a method that controls both errors.

The Factorized Directional Bandpass Method, Benchmark and Evaluation

In order to balance both errors we take the viewpoint that—loosely speaking—fingerprint images are highly determined by patterns that have frequencies only in a specific band of the Fourier spectrum (prior knowledge). Focusing on these frequencies occuring in true fingerprint images (FOTIs), we aim at the following goals:

  1. Equally preserving all FOTIs while attenuating all non-FOTIs.

  2. Removing all image artifacts in the FOTI spectrum, not due to the true fingerprint pattern.

  3. Returning a (smooth) texture image containing only FOTI features from the true fingerprint pattern.

  4. Morphological methods returning the ROI.

In order to meet these goals we have developed a factorized directional bandpass (FDB) segmentation method.

The FDB method

At the core of the FDB method is a classical Butterworth bandpass filter which guarantees Goal 1. Notably Goal 1 cannot fully be met by Gaussian based filtering methods such as the Gabor filter. Obviously, due to the Gaussian bell shaped curve, FOTIs would not be filtered alike. Because straightforward Fourier methods cannot cope with curvature (as could e.g. curved Gabor filters [2]) we perform separate filtering into a few isolated orientations only, via directional Hilbert transformations. The composite directional Hilbert Butterworth bandpass filter (DHBB) incorporates our prior knowledge about the range of possible values of ridge frequencies (between 1/3 and 1/25 pixels) or interridge distances (between 3 and 25 pixels) [2], assuming a sensor resolution of 500 DPI and that adult fingerprints are processed. In the case of adolescent fingerprints [3] or sensors with a different resolution, the images can be resized to achieve an age and sensor independent size—not only for the first segmentation step, but also for all later processing stages. Our parameters can be tuned to reach an optimal tradeoff between treating all realistic frequencies alike and avoiding Gibbs effects. Moreover we use a data friendly rectangular spectral shape of the bandpass filter employed which preserves the rectangular shape of the spatial image.

A second key ingredient is the factorization of the filter into two factors in the spectral domain, between which a thresholding operation is inserted. After preserving all FOTIs and removing all non-FOTIs in application of the first factor, all FOTI features not due to the true fingerprint pattern (which are usually less pronounced) are removed via a shrinkage operator: soft-thresholding. Note that albeit removing less pronounced FOTI features, thresholding introduces new unwanted high frequencies. These are removed, however, by application of the second factor, which also compensates for a possible phase shift due to the first factor, thus producing a smoothed image with pronounced FOTI features only.

At this stage, non-prominent FOTI features have been removed, not only outside the ROI, but also some due to true fingerprint features inside the ROI. In the final step, these “lost” regions are restored via morphological operations (convex hull after binarization and two-scale opening and closing).

The careful combination of the above ingredients in our proposed FDB method yields segmentation results far superior to existing segmentation methods. The procedure of the FDB method is illustrated in Fig 1.

Fig 1. Overview over the segmentation by the FDB method: In the analysis step, the original image (top row, left) is transformed into the Fourier domain (second column) and filtered by the first DHBB factor obtaining 16 directional subbands (third and fourth columns).

Fig 1

Next soft-thresholding is applied to remove spurious patterns (second row, third and fourth columns). In the synthesis step, the feature image (second column) is reconstructed from these subbands using the second DHBB factor. Finally, the feature image is binarized and the ROI is obtained by morphological operations. The estimated ROI (third row, left) is compared to manually marked ground truth segmentation (third row, right) in order to evaluate the segmentation performance.

Benchmark

In order to verify this claim, because of the lack of a suitable benchmark in the literature, we provide a manually marked ground truth segmentation for all 12 databases of FVC2000 [4], FVC2002 [5] and FVC2004 [6]. Each databases consists of 80 images for training and 800 images for testing. Overall this benchmark consists of 10560 marked segmentation images. This ground truth benchmark is made publicly available, so that other researchers can evaluate segmentation algorithms on it.

Evaluation against existing methods

We conduct a systematic performance comparison of widely used segmentation algorithms on this benchmark. In total, more than 100 methods for fingerprint segmentation can be found in literature. However, it remains unclear how these methods compare with each other in terms of segmentation performance and which methods can be considered as state-of-the-art. In order to remedy the current situation we chose four of the most often cited fingerprint segmentation methods and compared their performance: a method based on mean and variance of gray level intensities and the coherence of gradients as features and a neural network as a classifier [7], a method using Gabor filter bank responses [8], a Harris corner response based method [9] and an approach using local Fourier analysis [10].

Related Work

Early methods for fingerprint segmentation include Mehtre et al. [11] who segment an image based on histograms of local ridge orientation and in [12] additionally the gray-level variance is considered. A method proposed by Bazen and Gerez [7] uses the local mean and variance of gray-level intensities and the coherence of gradients as features and a neural network as a classifier. Similarly Chen et al. [13] use block based features including the mean and variance in combination with a linear classifier. Both methods perform morphology operations for postprocessing. A method by Shen et al. is based on Gabor filter bank responses of blocks [8]. In [2], all pixels are regarded as foreground for which a valid ridge frequency based on curved regions can be estimated. Wu et al. [9] proposed a Harris corner response based method and they apply Gabor responses for postprocessing. Wang et al. [14] proposed to use Gaussian-Hermite moments for fingerprint segmentation. The method of Zhu et al. [15] uses a gradient based orientation estimation as the main feature, and a neural network detects wrongly estimated orientation and classifies the corresponding blocks as background. Chikkerur et al. [10] applied local Fourier analysis for fingerprint image enhancement. The method performs implicitly fingerprint segmentation, orientation field and ridge frequency estimation. Further approaches for fingerprint enhancement in the Fourier domain include Sherlock et al. [16], Sutthiwichaiporn and Areekul [17] and Bartůněk et al. [1820]. Segmentation methods for latent fingerprints were proposed, see Zhang et al. [21], Nimkar and Mishra [22], Cao et al. [23], and the references therein. It would be of interest to see how these methods aiming at latent fingermarks perform on a benchmark of plain fingerprint images. Recently, Ferreira et al. [24] have proposed a method based on range filters and fuzzy C-means clustering for segmentation and binarization.

Setup of Paper

The paper is organized as follows: in the next section, we describe the proposed method beginning with the design of the DHBB filter for texture extraction in Section Filter Design for Fingerprint Segmentation. Subsequently, the extracted and denoised texture is utilized for estimating the segmentation as described in Fingerprint Segmentation which summarizes the FDB segmentation procedure. In Section Evaluation Benchmark and Results, the manually marked ground truth benchmark is introduced and applied for evaluating the segmentation performance of four widely used algorithms and for comparing them to the proposed FDB segmentation method. The results are discussed in Section Conclusions.

Fingerprint Segmentation by FDB Methods

Our segmentation method uses a filter transforming an input 2D image f(·)L2(R2) into a feature image

f˜(x):=l=0L-1xmZ2Tf(·),ϕlγ,n(·-m)L2pvcl[m],βdl[m]·ϕlγ,n(x-m). (1)

Due to our filter design, the L2 product above as well as all convolutions, integrals and sums are understood in the principal value sense

limϵ0y-mϵf(y)·ϕlγ,n(y-m)dy=f(·),ϕlγ,n(·-m)L2pv=(f*pvϕlγ,n,)(m).

Having clarified this, the symbol “pv” will be dropped in the following. At the core of Eq (1) is the DHBB filter conveyed by ϕlγ,n (l counts directions, n and γ are tuning parameters providing sharpness). In fact, we suitably factorize the filter conveyed by ϕlγ,n*ϕlγ,n, in the Fourier domain where ϕlγ,n,(x):=ϕlγ,n(-x) with the argument reversion operator “” and apply a thresholding procedure T “in the middle”. Underlying this factorization is a factorization of the bandpass filter involved. The precise filter design will be detailed in the following. Note that the directional Hilbert transform is also conveyed by a non-symmetric kernel. Reversing this transform (as well as the factor of the Butterworth) restores symmetry. It is inspired by the steerable wavelet [2527] and to some extend similar in spirit to the curvelet transform [28], [29] and the curved Gabor filters [2]. We deal with curvature by analyzing single directions l separately before the final synthesis.

Via factorization, possible phase shifts are compensated and unwanted frequencies introduced by the thresholding operator are eliminated, yielding a sparse smoothed feature image. This allows for easy binarization and segmentation via subsequent morphological methods, leading to the ROI.

Note that Eq (1) can be viewed as an analog to the projection operator in sampling theory with the analysis and synthesis steps (e.g. [30]). In this vein we have the following three steps:

  • Forward analysis (prediction): A first application of the argument reversed DHBB filter to a fingerprint image f corresponds to a number of directional selections in certain frequency bands of the fingerprint image giving cl[m] above.

  • Proximity operator (thresholding): In order to remove intermediate coefficients due to spurious patterns (cf. [31]) we perform soft tresholding on the filtered grey values yielding dl[m] above.

  • Backward synthesis: Subsequently we apply the filter (non-reversed) again giving f˜ assembled from all subbands. A numerical comparison to other synthesis methods, summation (corresponding to a naive reconstruction) and maximal response in the appendix Comparison of the Operator in the FDB Method with the Summation and Maximum Operators, shows the superiority of this smoothing step.

Due to the discrete nature of the image f[k]=f(x)x=kZ2, we work with the discrete version of f˜(x) at x=kZ2 in Eq (1).

Filter Design for Fingerprint Segmentation

The features of interest in a fingerprint image are repeated (curved) patterns which are concentrated in a particular range of frequencies after a Fourier transformation. In principle, the frequencies lower than these range’s limits correspond to homogeneous regions and those higher to small scale objects, i.e. noise, respectively. Taking this prior knowledge into account, we design an algorithm that captures these fingerprint patterns in different directional subbands in the frequency domain for extracting the texture.

In this section, we design angularpass and bandpass filters. The angularpass filter builds on iterates of the directional Hilbert transformation, a multidimensional generalization of the Hilbert transform called the Riesz transform. It can be represented via principal value convolution kernels. The bandpass filter builds on the Butterworth transform which can be represented directly via a convolution kernel. We follow here a standard technique designing a bandpass filter from a lowpass filter which has an equivalent representation in analog circuit design.

The nth Order Directional Hilbert Transform of a Butterworth Bandpass

Although a fingerprint image

k=[k1,k2]f[k],{-M,,M]×{-N,,N}{0,,255}

is a discrete signal observed over a discrete grid, M,NN we start our considerations with a signal

x=(x1,x2)f(x),D:=[-a,a]×[-b,b][0,1]

assuming values in a continuum a, b > 0. The frequency coordinates in the spectral domain will be denoted by ω=(ω1,ω2)R2.

As usual, the following operators are defined first for functions f in the Schwartz Space of rapidly-decaying and infinitely differentiable test functions:

S(Rd)=fC(Rd)supxRd(1+|x|m)dndxnf(x)<+,m,nZ+,

and continuously extended onto

L2(Rd)=fS(Rd)fL2=Rd|f(x)|2dx<+.

In our context we only need d = 1, 2. Further, we denote the Fourier and its inverse transformations by

F[f](ω)=Rdf(x)e-jω,xdx=f^(ω),F-1[f](ω)=1(2π)dRdf^(ω)ejω,xdx

where j denotes the imaginary unit with j2 = −1.

Butterworth bandpass

For γN and frequency bounds 0 < ωL < ωH, setting Δ = ωHωL, p2 = ωH ωL, the one-dimensional (d = 1) Butterworth bandpass transform is defined via

B[f](x)=F-1ω(ωΔ)2γ(ωΔ)2γ+(ω2-p2)2γ:=b^(ω)f^(ω)(x),

cf. [32]. It is easy to verify that b^(ω) tends to zero for ω → 0 and ω → ∞ and has unique maximum at the geometric mean p with value 1. In consequence, for high values of γ, this filter approximates the ideal filter

b^ideal(ω)=1ifωLωωH0else.

The ideal filter, however, suffers from the Gibbs effect. Letting t=(jω)2+p2jωΔ, we factorize the bandpass Butterworth as

b^2(ω)=1(-1)γ(t2)γ+1=1(-1)γk=1γ(t2-tk2)=k=1γ1t-tkH(t)·k=1γ1-t-tkH(-t),

with tk = eπj(γ + 2k − 1)/2γ (k = 1, …, γ), the negative squares of which representing the γ different complex roots of (−1). Then, with the below complex valued factor of 0b^2(ω)=B(jω)·B(-jω) called the transfer function,

H(t)=H(jω)2+p2jωΔ=k=1γΔ(jω)(jω)2-Δtk(jω)+p2:=B(jω),

we use the approximation: jω=logejω2ejω-1ejω+1 to obtain

B(jω)k=1γ2Δ(e2jω-1)4+p2-2Δtke2jω+2p2-8ejω+4+p2+2Δtk:=Bγ(ejω).

This approximation is often called the bilinear transform, which turns out to reduce the frequency bandwidth of interest, cf. Fig 2.

Fig 2.

Fig 2

Image (a) displays a 1D Butterworth bandpass filter (blue) and its approximation (red). Image (b) shows the 2D Butterworth bandpass filter g^γ(ω) at n = 20, θ = 0, γ = 2, and the corresponding DHBB filter in the Fourier and spatial domains (c, d) for the approximation by the bilinear transform. Image (e) visualizes the 2D version of the original filter and the corresponding DHBB filter (f, g).

The 1D filter Bγ(e) is then generalized to a 2D domain. The McClellan transform [3336], would be one favorable method. Also, recently, bandpass filtering with a radial filter in the Fourier domain has been proposed by [37], [38] and [39] et al. for enhancing fingerprint images. However, for a simpler reconstruction of 2D filter and a data-friendly alternative to the polar tiling of the frequency plane, a Cartesian array is used instead (see [28], [29], [40], [41]).

Thus, on a rectangular domain D = [−a, a] × [−b, b] with common cuttoff frequencies 0 < ωL < ωH and the two characteristic functions

χh(ω1,ω2):=1ifb|ω1|a|ω2|0else,χv(ω1,ω2):=1ifb|ω1|a|ω2|0else

(see Fig 3), define

g^γ(ω1,ω2)=Bγ(ejω1)χh(ω1,ω2)+Bγ(ejω2)χv(ω1,ω2) (2)

as the spectrum of our two-dimensional Butterworth filter gγ(x). Note that since g^γ(ω)L2(R2), there is a well defined gγ(x)=F-1[g^γ(ω)](x).

Fig 3. The indicator functions in the horizontal direction (a) and vertical direction (b).

Fig 3

Fig 4(b) and 4(c) show an example of the 1D and 2D Butterworth bandpass filters.

Fig 4.

Fig 4

Image (a) displays the angularpass filter h^ln(ω) with θ = 13π/16, n = 20. Images (b-c) show the 1D and 2D Butterworth bandpass filters g^γ(ω) and g^γ(ω) with ωL = 0.3, ωH = 1, γ = 2, (d) the spectrum of the DHBB filter ϕ^lγ,n(ω). Images (e-f) visualize the real and imaginary part of the DHBB filter ϕlγ,n(x). Images (g-h) display the squared magnitude of the spectrum of the DHBB in the frequency and spatial domains which acts somewhat like a Gabor filter.

n-th order directional Hilbert transformations

For more detail on the Hilbert transform H and the Riesz transform R, we refer the reader to the literature for an in-depth discussion [25], [27], [42], [43], [44], [4547], [48], and [49].

Consider a vector uRd and set and compute, respectively, for xRd

R[f](x):=F-1ω-jωω:=h^(ω)f^(ω)(x) (3)
Hu[f](x):=u,R[f](x)=F-1ω-ju,ωω:=h^u(ω)f^(ω)(x)
HuHun-times[f](x)=Hun[f](x)=F-1ω(-j)nu,ωnωnh^un(ω)f^(ω)(x),nN (4)

The first line Eq (3) called the Riesz transform has a representation as a principal value integral

R[f](x)=(f*pvh)(x)=limϵ0y-xϵf(x-y)h(y)dy

where

h(y):=yyd+1Γd+12πd+12ford>11πyford=1

Setting hu(y) = 〈u, h(y)〉 and hun(y)=(hu*pv*pvhu)(y) we haven for the third line Eq (4) called the n-order directional Hilbert transform that

Hun[f](x)=f*pvhun(x).

Since -1u,ωω1 and high powers preserve the values near ±1 while forcing all other values in (−1, 1) towards 0, this filter gives roughly the same result as an inverse Fourier transform of a convolution of the signal’s Fourier transform with

Aα,u(ω)=1if|u,ω|ωcosα0else

for small α > 0. The directional Hilbert transform, however, suffers less from a Gibbs effect than this sharp cutoff filter.

In 2D, the direction vector is u = [cos(θ), sin(θ)]T with the discretized θ=πlL[0,π) and l = 0, 1, …, L − 1, where LN is the total number of orientation. Rewrite the impulse response of the n-thN order directional Hilbert transform in Eq (4) as

h^un(ω1,ω2)=-jcostan2-1ω2ω1-πlLn:=h^ln(ω1,ω2). (5)

Putting together (Eqs (2), (4) and (5)), for a fixed bandpass ωL < ωH and L directional subbands we have thus the DHBB filter of order γ, n:

Hun[gγ](x)=F-1ωh^ln(ω)·g^γ(ω):=ϕ^lγ,n(ω)(x):=ϕlγ,n(x). (6)

Thresholding

For given β > 0, soft-thresholding is defined as follows

xT(x,β)=x|x|·max|x|-β,0. (7)

Thus, the thresholded coefficients are dl [m] = T{cl [m], β}. Note that dl[m] is a solution of the 1-shrinkage minimization problem

minuβu1+12u-cl22

yielding soft-thresholding (cf. [50]). Fig 5 visualizes the effect of the soft-thresholding and the comparison with the others (such as: hard [50], semi-soft [51] and nonlinear [52] thresholding operators).

Fig 5. Four typical thresholding functions (red: hard, black: soft, green: semi-soft, magenta: nonlinear) are compared (top left).

Fig 5

The following six pairs show an image and the visualization of the corresponding 1D cross section along the red line. F.l.t.r and top to bottom: the original image f[k], the coefficient cl[k] and the thresholded coefficients dl[k] for the soft, hard, semisoft and nonlinear thresholding operators. Comparing the four cross sections in the bottom row, we observe that soft-thresholding achieves the sparsest solution.

Fingerprint Segmentation

After having designed the FDB filter, let us now ponder on parameter selection, image binarization and morphological processing.

Parameter Choice for Texture Extraction

A fingerprint image will be rescaled such that its oscillation pattern stays in a specific range in the Fourier domain, the coordinates of which are ωi = [−π, π], i ∈ {1, 2}. For choosing the cutoff frequencies ωL and ωH, we incorporate our prior knowledge about adult fingerprint images at resolution of 500 DPI: Valid interridge distances remain in a known range approximately from 3 to 25 pixels [2]. This corresponds exactly to ωH = 1 as a limit for high frequencies. A limit of ωL = 0.3 for low frequencies of the Butterworth bandpass filter corresponds to an interridge distance of about 12 pixels. The range |ωi| ∈ [ωH, π] contains the small scale objects which are considered as noise. The range |ωi| ∈ [0, ωL] contains the low frequency objects, corresponding to homogeneous regions.

The number of directions L in and the order n of the directional Hilbert transform involves a tradeoff between the following effects. We observe that with increased order n the filter’s shape becomes thinner in the Fourier domain. Although this sparsity smooths the texture image in the spatial domain, in order to fully cover all FOTIs, L needs to grow with n. However, a disadvantage of choosing large n and L is that errors occur on the boundary due to the over-smoothing effect as illustrated in Fig 6 (o).

Fig 6. Angular bandpass h^ln(ω) at θ=7π16,γ=3 and different orders n ∈ {3, 20, 100} and their responses (last row).

Fig 6

The next parameter to select is the order of the Butterworth filter γ. An illustration of the filter for different orders γ ∈ {1, 2, 3, 10} and with cutoff frequencies ωL = 0.3 and ωH = 1 is shown in Fig 7, its bilinear approximation in Fig 2. As γ increases the filter becomes sharper. For very large values of γ, it approaches the ideal filter which is known to cause the unfavorable Gibbs effect.

Fig 7. Butterworth bandpass filter g^γ(ω) at ωL = 0.3, ωH = 1 and different γ, angular bandpass filter with n=20,L=16,θ=5π16, and their responses.

Fig 7

1st row: 1D Butterworth, 2nd row: 2D Butterworth, 3rd row: ϕ^lγ,n(ω), 4th row: |ϕlγ,n(x)|, 5th row: (ϕlγ,n*ϕlγ,n,)(x), 6th row: their responses.

The thresholding value β separates large coefficients corresponding to the fingerprint pattern (FOTIs) (which are slightly attenuated due to soft-thresholding) from small coefficients corresponding to non-FOTIS and FOTIs which are not features due to the fingerprint pattern (these are eliminated). On the one hand, if β is chosen too large, more prominent parts of true fingerprint tend to be removed. On the other hand, if β is chosen too small, not all all unwanted features (as above) are removed which may cause segmentation errors.

In order to find good trade-offs, as described above, n, L, γ and β are trained as described in Section Experimental Results. In fact, since different fingerprint sensors have different properties, β is adaptively adjusted to the intensity of coefficients in all subbands as

β=C·maxl,m{cl[m]}. (8)

Thus, instead of β, C is trained for each sensor.

Texture Binarization

In the first step, the texture is decomposed by the operator Eq (1) to obtain the reconstructed image f˜[k]. Then, f˜[k] is binarized using an adaptive threshold adjusted to the intensity of f˜[k]. Thus, the threshold is chosen as C·maxk(f˜[k]), with C from Eq (8). If f˜[k] is larger than this threshold, it will be set to 1 (foreground), otherwise, it is set to 0 (background) as illustrated in Fig 1.

f˜bin[k]=1,f˜[k]C·maxk(f˜[k])0,else,kΩ

Morphological Processing

In this final phase, we apply mathematical morphology (see Chapter 13 in [53]), to decide for each pixel whether it belongs to the foreground or background. Firstly, at each pixel f˜bin[k1,k2]{0,1}, we build an s × s block centered at (k1, k2) and 8 neighboring blocks (cf. Fig 8). Then, for each block, we count the white pixels and check whether their number exceeds the threshold s2t with another parameter t > 0. If at least b blocks are above threshold, the pixel [k1, k2] is considered as foreground.

f˜dilate[k1,k2]=1,#B[k1+m,k2+m]s2t,m{-s,0,s}b0,else (9)
Fig 8. The morphological element.

Fig 8

Then, the largest connected white pixel component is selected by a region filling method. Its convex hull is then the ROI. For better visualization we have inverted white and black, i.e. display the background by white pixels and the ROI by black pixels, cf. Fig 1. Fig 9 illustrates the effect of the morphological operator.

Fig 9.

Fig 9

The ground truth segmentation (a) and the binarized texture image (b) for an example fingerprint. Applying a standard morphology operation like closing (dilation followed by erosion) instead of the proposed method connects in this example the white fingerprint texture with structure noise close to the margin of the texture and the result is a defective segmentation (c). The proposed morphology avoids this undesired effect by considering neighborhoods on two scales: cells of size s × s pixels and blocks of 3 × 3 cells.

Evaluation Benchmark and Results

The databases of FVC2000, 2002 and 2004 [46] are publicly available and established benchmarks for measuring the verification performance of algorithms for image enhancement and fingerprint matching. Each competition comprises four databases: three of which contain real fingerprints acquired by different sensors and a database of synthetically generated images (DB 4 in each competition).

It has recently been shown that real and synthetic fingerprints can be discriminated with very high accuracy using minutiae histograms (MHs) [54]. More specifically, by computing the MH for a minutiae template and then computing the earth mover’s distance (EMD) [55] between the template MH and the mean MHs for a set of real and synthetic fingerprints. Classification is simply performed by choosing the class with the smaller EMD.

The nine databases containing real fingerprints have been obtained by nine different sensors and have different properties. The fingerprint image quality ranges from good quality images (especially FVC2002 DB1 and DB2) to low quality images which are more challenging to process (e.g. the databases of FVC2004). Some aspects of image quality concern both the segmentation step and the overall verification process, other aspects pose problems only for later stages of the fingerprint verification procedure, but have no influence on the segmentation accuracy.

Aspects of fingerprint image quality which complicate the segmentation:

  • dryness or wetness of the finger

  • a ghost fingerprint on the sensor surface

  • small scale noise

  • large scale structure noise

  • image artifacts e.g. caused by reconstructing a swipe sensor image

  • scars or creases interrupting the fingerprint pattern

Aspects of fingerprint image quality which make an accurate verification more difficult, but do not have any influence on the fingerprint segmentation step:

  • distortion, nonlinear deformation of the finger

  • small overlap area between two imprints

Each of the 12 databases contains 110 fingers with 8 impressions per finger. The training set consists of 10 fingers (80 images) and the test set contains 100 fingers (800 images). In total there are 10560 fingerprint images giving 10560 marked ground truth segmentations for training and testing.

Experimental Results

Segmentation Performance Evaluation

Let N1 and N2 be the width and height of image f[k] in pixels. Let Mf be number of pixels which are marked as foreground by human expert and estimated as background by an algorithm (missed/misclassified foreground). Let Mb be number of pixels which are marked as background by human expert and estimated as foreground by an algorithm (missed/misclassified background). The average total error per image is defined as

Err=Mf+MbN1×N2. (10)

The average error over 80 training images is basis for the parameter selection. In Table 1, we report the average error over all other 800 test images for each database and for each algorithms.

Table 1. Error rates (average percentage of misclassified pixels averaged over 800 test images per database) computed using the manually marked ground truth segmentation and the estimated segmentation by these methods: a Gabor filter bank (GFB) response based method by Shen et al. [8], a Harris corner response (HCR) based approach by Wu et al. [9], a method by Bazen and Gerez using local gray-level mean, variance and gradient coherence (MVC) as features [7], a method applying short time Fourier transforms (STFT) by Chikkerur et al. [10] and the proposed method based on the factorized directional bandpass (FDB).
FVC DB GFB [8] HCR [9] MVC [7] STFT [10] FDB
2000 1 13.26 11.15 10.01 16.70 5.51
2 10.27 6.25 12.31 8.88 3.55
3 10.63 7.80 7.45 6.44 2.86
4 5.17 3.23 9.74 7.19 2.31
2002 1 5.07 3.71 4.59 5.49 2.39
2 7.76 5.72 4.32 6.27 2.91
3 9.60 4.71 5.29 5.13 3.35
4 7.67 6.85 6.12 7.70 4.49
2004 1 5.00 2.26 2.22 2.65 1.40
2 11.18 7.54 8.06 9.89 4.90
3 8.37 4.96 3.42 9.35 3.14
4 5.96 5.15 4.58 5.18 2.79
Avg. 8.33 5.78 6.51 7.57 3.30

Parameter Selection

Experiments were carried out on all 12 databases and are reported in Table 1. For each method listed in Table 1, the required parameters were trained on each of the 12 training sets: the choice of the threshold values for the Gabor filter bank based approach by Shen et al. [8], and the threshold values for the Harris corner response based method by Wu et al. [9]. The parameters of the method by Bazen and Gerez are chosen as described in [7]: the window size of the morphology operator and the weights of the perceptron which are trained in 104 iterations due to the large number of pixels in the training database. For the method of Chikkerur et al., we used the energy image computed by the implementation of Chikkerur, performed Otsu thresholding and mathematical morphology as explained in [53].

For the proposed FDB method, the involved parameters are summarized in Table 2 and the values of the learned parameters are reported in Table 3. Also, the mirror boundary condition with size 15 pixels is used in order to avoid boundary effects. In a reasonable amount of time, a number of conceivable parameter combinations were evaluated on the training set. The choice of these parameters balances the smoothing properties of the proposed filter attempting to avoid both under-smoothing and over-smoothing. The response of a fingerprint image by FDB is illustrated in Fig 10.

Table 2. Overview over all parameters for the factorized directional bandpass (FDB) method for fingerprint segmentation.
Parameters Description
C a constant for selecting the threshold β in Eq (8) which removes small coefficients corresponding to noise.
n the order of the directional Hilbert transform which corresponds to the angularpass filter in Eq (4).
L the number of orientations in the angularpass filter in Eq (4).
γ the order of the Butterworth bandpass filter in Eq (2).
s the window size of the block in the postprocessing step in Eq (9).
t a constant for selecting the morphology threshold T in Eq (9).
b the number of the neighboring blocks in Eq (9).
Table 3. Overview over the parameters learned on the training set.

The other four parameters are n = 20, L = 16, s = 9 and b = 6 for all databases.

FVC DB C γ t
2000 1 0.06 4 5
2 0.07 2 5
3 0.06 4 4
4 0.03 1 5
2002 1 0.04 1 4
2 0.05 1 7
3 0.09 1 5
4 0.03 1 6
2004 1 0.04 1 7
2 0.08 2 5
3 0.07 1 6
4 0.05 1 5
Fig 10. Visualization of the coefficients in the 16 subbands of the DHBB filter forn = 20, γ = 3, ωL = 0.3, ωH = 1.

Fig 10

This systematic comparison of fingerprint segmentation methods clearly shows that the factorized directional bandpass method (FDB) outperforms the other four widely used segmentation methods on all 12 databases. An overview of visualized segmentation results by the FDB method is given in Fig 11. A few challenging examples for which the FDB method produces a flawed segmentation are depicted in Fig 12. Moreover, a comparison of all five segmentation methods and their main features for five example images are shown in Figs 13 to 17.

Fig 11. Segmented fingerprint images and the corresponding reconstructed texture images by the FDB method for FVC2000 (first and second row), FVC2002 (third and fourth row) and FVC2004 (fifth and sixth row).

Fig 11

Columns f.l.t.r correspond to DB1 to DB4.

Fig 12. Examples of incorrectly segmented fingerprint images due to: (a) a ghost fingerprint on the sensor surface, (b) dryness of the finger, (c) texture artifacts in the reconstructed image, (d) wetness of the finger.

Fig 12

The first row shows the segmentation obtained by the FDB method, the second row displays the reconstructed image and the last row visualizes the manually marked ground truth segmentation.

Fig 13. Segmented fingerprint images and their features of different methods for FVC2002_DB3_IM_15_1.

Fig 13

(a) ground truth; (b, g) FDB, (c, h) Gabor, (d, i) Harris, (e, j, k, l) Mean-Variance-Coherence, (f, m) STFT.

Fig 17. Segmented fingerprint images and their features of different methods for FVC2004_DB2_IM_65_7.

Fig 17

(a) ground truth; (b, g) FDB, (c, h) Gabor, (d, i) Harris, (e, j, k, l) Mean-Variance-Coherence, (f, m) STFT.

Fig 14. Segmented fingerprint images and their features of different methods for FVC2004_DB1_IM_24_7.

Fig 14

(a) ground truth; (b, g) FDB, (c, h) Gabor, (d, i) Harris, (e, j, k, l) Mean-Variance-Coherence, (f, m) STFT.

Fig 15. Segmented fingerprint images and their features of different methods for FVC2000_DB3_IM_17_3.

Fig 15

(a) ground truth; (b, g) FDB, (c, h) Gabor, (d, i) Harris, (e, j, k, l) Mean-Variance-Coherence, (f, m) STFT.

Fig 16. Segmented fingerprint images and their features of different methods for FVC2004_DB2_IM_56_8.

Fig 16

(a) ground truth; (b, g) FDB, (c, h) Gabor, (d, i) Harris, (e, j, k, l) Mean-Variance-Coherence, (f, m) STFT.

Conclusions

In this paper, we designed a filter specifically for fingerprints which is based on the directional Hilbert transform of Butterworth bandpass filters. A systematic comparison with four widely used fingerprint segmentation showed that the proposed FDB method outperforms these methods on all 12 FVC databases using manually marked ground truth segmentation for the performance evaluation. The proposed FDB method for fingerprint segmentation can be combined with all methods for orientation field estimation like e.g. the line sensor method [56] or by a global model based on quadratic differentials [57] followed by liveness detection [58] or fingerprint image enhancement [2, 59]. It can also be used in combination with alternative approaches, e.g. as a preprocessing step for locally adaptive fingerprint enhancement in the Fourier domain as proposed by Bartůněk et al. [20] or before applying structure tensor derived symmetry features for enhancement and minutiae extraction proposed by Fronthaler et al. [60].

Notably, the filter ϕln,γ*ϕln,γ, is similar to the Gabor filter which could have been used instead of the DHBB filter. Similarly, Bessel or Chebbychev transforms as well as B-splines as generalizations ([61]) could replace the Butterworth. We expect, however, for reasons elaborated, relying on the DHBB filter gives superior segmentation results.

The manually marked ground truth benchmark is available for download at http://dx.doi.org/10.6084/m9.figshare.1294209 and the Matlab implementation of the FDB method is available for download at http://dx.doi.org/10.6084/m9.figshare.1294210.

In doing so, we would like to facilitate the reproducibility of the presented results and promote the comparability of fingerprint segmentation methods. Recently, this implementation of the FDB method has been applied to improve the performance of fingerprint liveness detection by the convolution comparison patterns [62] and fingerprint alteration detection [63]. The manually marked benchmark has been used by Thai and Gottschlich [64] and by Bartůněk [65] for evaluating a new fingerprint segmentation methods. The G3PD method [64] follows a variational approach to decompose a fingerprint image into three parts and obtains the ROI based on the texture component. The further advanced and more general DG3PD method [66] has also been applied to latent fingerprint segmentation. The method by Bartůněk [65] relies on normalisation and local kurtosis estimation as a novel feature for segmentation.

Supplementary Appendix

Comparison of the Operator in the FDB Method with the Summation and Maximum Operators

We briefly illustrate the differences between the proposed FDB filter Eq (1) and the maximum and summation operators for the coefficients in all directional subbands. Fig 18 compares the results of these operators for a low-quality and a good quality example. The functions are described as follows

Fig 18. Comparison of five image reconstruction strategies and their effect on the resulting segmentation.

Fig 18

1st, 2nd columns: segmented images (error in percent) and reconstructed images for a low-quality image and 3rd, 4th columns for a good quality image. 1st row: the proposed operator. 2nd, 3rd rows: maximum operator without and with the shrinkage operator Eq (7), respectively. 4th, 5th rows: summation operator without and with the shrinkage operator Eq (7), respectively.

  • The maximum operator without and with the shrinkage operator Eq (7) (depicted in the second and third row in Fig 18)
    f˜[k]=maxlcl[k]·(cl[k]>0)+minlcl[k]·(cl[k]<0)without(7)maxldl[k]·(dl[k]>0)+minldl[k]·(dl[k]<0)with(7),
    with l = 0, 1, …, L − 1.
  • The summation operator without and with the shrinkage operator Eq (7) (displayed in the fourth and fifth row in Fig 18)
    f˜[k]=l=0L-1cl[k]without(7)l=0L-1dl[k]with(7).

Acknowledgments

The authors gratefully acknowledge support by the German Research Foundation (DFG) and the Open Access Publication Funds of the University of Göttingen. D.H Thai is supported by the National Science Foundation under Grant DMS-1127914 to the Statistical and Applied Mathematical Sciences Institute. S. Huckemann and C. Gottschlich gratefully acknowledge the support of the Felix-Bernstein-Institute for Mathematical Statistics in the Biosciences and the Niedersachsen Vorab of the Volkswagen Foundation. The authors would like to thank Benjamin Eltzner, Axel Munk, Gerlind Plonka-Hoch and Yvo Pokern for their valuable comments.

Data Availability

The manually marked ground truth benchmark is available for download at http://dx.doi.org/10.6084/m9.figshare.1294209 and the Matlab implementation of the FDB method is available for download at http://dx.doi.org/10.6084/m9.figshare.1294210.

Funding Statement

The authors gratefully acknowledge support by the German Research Foundation (DFG) and the Open Access Publication Funds of the University of Gottingen. D.H Thai is supported by the National Science Foundation under Grant DMS-1127914 to the Statistical and Applied Mathematical Sciences Institute. S. Huckemann and C. Gottschlich gratefully acknowledge the support of the Felix-Bernstein-Institute for Mathematical Statistics in the Biosciences and the Niedersachsen Vorab of the Volkswagen Foundation. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

References

  • 1. Maltoni D, Maio D, Jain AK, Prabhakar S. Handbook of Fingerprint Recognition. London, U.K.: Springer; 2009. [Google Scholar]
  • 2. Gottschlich C. Curved-Region-Based Ridge Frequency Estimation and Curved Gabor Filters for Fingerprint Image Enhancement. IEEE Transactions on Image Processing. 2012. April;21(4):2220–2227. 10.1109/TIP.2011.2170696 [DOI] [PubMed] [Google Scholar]
  • 3. Gottschlich C, Hotz T, Lorenz R, Bernhardt S, Hantschel M, Munk A. Modeling the Growth of Fingerprints Improves Matching for Adolescents. IEEE Transactions on Information Forensics and Security. 2011. September;6(3):1165–1169. 10.1109/TIFS.2011.2143406 [DOI] [Google Scholar]
  • 4. Maio D, Maltoni D, Capelli R, Wayman JL, Jain AK. FVC2000: Fingerprint Verification Competition. IEEE Transactions on Pattern Analysis and Machine Intelligence. 2002. March;24(3):402–412. 10.1109/34.990140 [DOI] [PubMed] [Google Scholar]
  • 5.Maio, D, Maltoni, D, Capelli, R, Wayman, JL, Jain, AK. FVC2002: Second Fingerprint Verification Competition. In: Proc. ICPR; 2002. p. 811–814.
  • 6.Maio D, Maltoni D, Capelli R, Wayman JL, Jain AK. FVC2004: Third Fingerprint Verification Competition. In: Proc. ICBA. Hong Kong; 2004. p. 1–7.
  • 7.Bazen AM, Gerez SH. Segmentation of Fingerprint Images. In: Proc. ProRISC. Veldhoven, The Netherlands; 2001. p. 276–280.
  • 8.Shen LL, Kot A, Koo WM. Quality measures of fingerprint images. In: Proc. AVBPA. Halmstad, Sweden; 2001. p. 266–271.
  • 9.Wu C, Tulyakov S, Govindaraju V. Robust point-based feature fingerprint segmentation algorithm. In: Proc. ICB 2007. Seoul, Korea; 2007. p. 1095–1103.
  • 10. Chikkerur S, Cartwright A, Govindaraju V. Fingerprint image enhancement using STFT analysis. Pattern Recognition. 2007;40(1):198–211. 10.1016/j.patcog.2006.05.036 [DOI] [Google Scholar]
  • 11. Mehtre BM, Murthy NN, Kapoor S, Chatterjee B. Segmentation of fingerprint images using the directional image. Pattern Recognition. 1987. August;20(4):429–435. 10.1016/0031-3203(87)90069-0 [DOI] [Google Scholar]
  • 12. Mehtre BM, Chatterjee B. Segmentation of fingerprint images—a composite method. Pattern Recognition. 1989. August;22(4):381–385. 10.1016/0031-3203(89)90047-2 [DOI] [Google Scholar]
  • 13. Chen X, Tian J, Cheng J, Yang X. Segmentation of Fingerprint Images Using Linear Classifier. EURASIP Journal on Applied Signal Processing. 2004;2004(4):480–494. 10.1155/S1110865704309194 [DOI] [Google Scholar]
  • 14.Wang L, Suo H, Dai M. Fingerprint Image Segmentation Based on Gaussian-Hermite Moments. In: Proc. ADMA. Wuhan, China; 2005. p. 446–454.
  • 15. Zhu E, Yin J, Hu C, Zhang G. A systematic method for fingerprint ridge orientation estimation and image segmentation. Pattern Recognition. 2006. August;39(8):1452–1472. 10.1016/j.patcog.2006.03.001 [DOI] [Google Scholar]
  • 16. Sherlock BG, Monro DM, Millard K. Fingerprint enhancement by directional Fourier filtering. IEE Proc Vision, Image and Signal Processing. 1994. April;141(2):87–94. 10.1049/ip-vis:19949924 [DOI] [Google Scholar]
  • 17. Sutthiwichaiporn P, Areekul V. Adaptive boosted spectral filtering for progressive fingerprint enhancement. Pattern Recognition. 2013. September;46(9):2465–2486. 10.1016/j.patcog.2013.02.002 [DOI] [Google Scholar]
  • 18.Bartůněk JS, Nilsson M, Nordberg J, Claesson I. Adaptive Fingerprint Binarization by Frequency Domain Analysis. In: Proc. ACSSC. Pacific Grove, CA, USA; 2006. p. 598–602.
  • 19.Bartůněk JS, Nilsson M, Nordberg J, Claesson I. Improved Adaptive Fingerprint Binarization. In: Proc. CISP. Sanya, China; 2008. p. 756–760.
  • 20. Bartůněk JS, Nilsson M, Sällberg B, Claesson I. Adaptive Fingerprint Image Enhancement With Emphasis on Preprocessing of Data. IEEE Transactions on Image Processing. 2013. February;22(2):644–656. 10.1109/TIP.2012.2220373 [DOI] [PubMed] [Google Scholar]
  • 21. Zhang J, Lai R, Kuo CCJ. Adaptive Directional Total-Variation Model for Latent Fingerprint Segmentation. IEEE Transactions on Information Forensics and Security. 2013. August;8(8):1261–1273. 10.1109/TIFS.2013.2267491 [DOI] [Google Scholar]
  • 22. Nimkar R, Mishra A. Fingerprint Segmentation Algorithms: A Literature Review. International Journal of Computer Applications. 2014. June;95(5):20–24. [Google Scholar]
  • 23. Cao K, Liu E, Jain AK. Segmentation and Enhancement of Latent Fingerprints: A Coarse to Fine Ridge Structure Dictionary. IEEE Transactions on Pattern Analysis and Machine Intelligence. 2014. September;36(9):1847–1859. 10.1109/TPAMI.2014.2302450 [DOI] [PubMed] [Google Scholar]
  • 24.Ferreira PM, Sequeira AF, Rebelo A. A Fuzzy C-Means Algorithm for Fingerprint Segmentation. In: Proc. Iberian Conference on Pattern Recognition and Image Analysis. Santiago de Compostela, Spain; 2015. p. 245–252.
  • 25. Unser M, Ville DVD. Wavelet Steerability and the Higher-Order Riesz Transform. IEEE Transactions on Image Processing. 2010. March;19(3):636–652. 10.1109/TIP.2009.2038832 [DOI] [PubMed] [Google Scholar]
  • 26. Held S, Storath M, Massopust P, Forster B. Steerable Wavelet Frames Based on the Riesz Transform. IEEE Transactions on Image Processing. 2010. March;19(3):653–667. 10.1109/TIP.2009.2036713 [DOI] [PubMed] [Google Scholar]
  • 27. Unser M, Chenouard N, Ville DVD. Steerable Pyramids and Tight Wavelet Frames in L2(Rd). IEEE Transactions on Image Processing. 2011. October;20(10):2705–2721. 10.1109/TIP.2011.2138147 [DOI] [PubMed] [Google Scholar]
  • 28. Ma J, Plonka G. The Curvelet Transform. IEEE Signal Processing Magazin. 2010. March;27(2):118–133. 10.1109/MSP.2009.935453 [DOI] [Google Scholar]
  • 29. Candès E, Demanet L, Donoho D, Ying L. Fast Discrete Curvelet Transforms. Multiscale Model Simul. 2006. September;5(3):861–899. 10.1137/05064182X [DOI] [Google Scholar]
  • 30. Unser M. Sampling—50 Years After Shannon. Proceedings of the IEEE. 2000. April;88(4):569–587. 10.1109/5.843002 [DOI] [Google Scholar]
  • 31. Haltmeier M, Munk A. Extreme value analysis of empirical frame coefficients and implications for denoising by soft-thresholding. Applied and Computational Harmonic Analysis. 2014. May;36(3):434–460. 10.1016/j.acha.2013.07.004 [DOI] [Google Scholar]
  • 32. Proakis JG, Manolakis DK. Digital Signal Processing: Principles, Algorithms, and Applications. Upper Saddle River, NJ, USA: Prentice Hall; 2007. [Google Scholar]
  • 33.McClellan JH. The design of two-dimensional digital filters by transformations. In: Proc. APCIS. Princeton, NJ, USA; 1973. p. 247–251.
  • 34. Mersereau RM, Mecklenbräuker WFG, T F Quatieri J. McClellan transformations for two-dimensional digital filtering: I—design. IEEE Transactions on Circuits and Systems. 1976. July;23(7):405–414. 10.1109/TCS.1976.1084236 [DOI] [Google Scholar]
  • 35. Mecklenbräuker WFG, Mersereau RM. McClellan transformations for two-dimensional digital filtering: II—Implementation. IEEE Transactions on Circuits and Systems. 1976. July;23(7):414–422. 10.1109/TCS.1976.1084237 [DOI] [Google Scholar]
  • 36. Tseng CC. Design of two-dimensional FIR digital filters by McClellan transform and quadratic programming. IEE Proceedings—Vision, Image and Signal Processing. 2001. October;148(5):325–331. 10.1049/ip-vis:20010410 [DOI] [Google Scholar]
  • 37. Kočevar M, Kotnik B, Chowdhury A, Kačič Z. Real-time fingerprint image enhancement with a two-stage algorithm and block-local normalization. Journal of Real-Time Image Processing. 2014. July;p. 1–10. [Google Scholar]
  • 38. Ghafoor M, Taj IA, Ahmad W, Jafri NM. Efficient 2-fold contextual filtering approach for fingerprint enhancement. IET Image Processing. 2014. July;8(7):417–425. 10.1049/iet-ipr.2013.0528 [DOI] [Google Scholar]
  • 39. Yang J, Xiong N, Vasilakos AV. Two-Stage Enhancement Scheme for Low-Quality Fingerprint Images by Learning From the Images. IEEE Transactions on Human-Machine Systems. 2013. March;43(2):235–248. 10.1109/TSMCC.2011.2174049 [DOI] [Google Scholar]
  • 40. Yi S, Labate D, Easley GR, Krim H. A Shearlet Approach to Edge Analysis and Detection. IEEE Transactions on Image Processing. 2009. May;18(5):929–941. 10.1109/TIP.2009.2013082 [DOI] [PubMed] [Google Scholar]
  • 41. Do MN, Vetterli M. The contourlet transform: An efficient directional multiresolution image representation. IEEE Transactions on Image Processing. 2005. December;14(12):2091–2106. 10.1109/TIP.2005.859376 [DOI] [PubMed] [Google Scholar]
  • 42. Chaudhury KN, Unser M. Construction of Hilbert Transform Pairs of Wavelet Bases and Gabor-like Transforms. IEEE Transactions on Signal Processing. 2009. September;57(9):3411–3425. 10.1109/TSP.2009.2020767 [DOI] [Google Scholar]
  • 43. Felsberg M, Sommer G. The Monogenic Signal. IEEE Transactions on Signal Processing. 2001. December;49(12):3136–3144. 10.1109/78.969520 [DOI] [Google Scholar]
  • 44. Hahn SL. Hilbert transforms in signal processing. Boston, MA, USA: Artech House; 1996. [Google Scholar]
  • 45. Larkin KG, Bone DJ, Oldfield MA. Natural Demodulation of Two-dimensional Fringe Patterns. I. General Background of the Spiral Phase Quadrature Transform. J Opt Soc Am A. 2001. August;18(8):1862–1870. 10.1364/JOSAA.18.001871 [DOI] [PubMed] [Google Scholar]
  • 46. Larkin KG. Natural Demodulation of Two-dimensional Fringe Patterns. II. Stationary phase analysis of the spiral phase quadrature transform. J Opt Soc Am A. 2001. August;18(8):1871–1881. 10.1364/JOSAA.18.001871 [DOI] [PubMed] [Google Scholar]
  • 47. Larkin KG, Fletcher PA. A Coherent Framework for Fingerprint Analysis: Are Fingerprints Holograms? Optics Express. 2007;15(14):8667–8677. 10.1364/OE.15.008667 [DOI] [PubMed] [Google Scholar]
  • 48. Unser M, Sage D, Ville DVD. Multiresolution Monogenic Signal Analysis Using the Riesz-Laplace Wavelet Transform. IEEE Transactions on Image Processing. 2009. November;18(11):2402–2418. 10.1109/TIP.2009.2027628 [DOI] [PubMed] [Google Scholar]
  • 49. Häuser S, Heise B, Steidl G. Linearized Riesz Transform and Quasi-Monogenic Shearlets. International Journal of Wavelets, Multiresolution and Information Processing. 2014. May;12(3). [Google Scholar]
  • 50. Donoho DL, Johnstone JM. Ideal spatial adaptation by wavelet shrinkage. Biometrika. 1994. August;81(3):425–455. 10.1093/biomet/81.3.425 [DOI] [Google Scholar]
  • 51. Sanam TF, Shahnaz C. A semisoft thresholding method based on Teager energy operation on wavelet packet coefficients for enhancing noisy speech. EURASIP Journal on Audio, Speech, and Music Processing. 2013. November;2013(1):1–15. 10.1186/1687-4722-2013-25 [DOI] [Google Scholar]
  • 52. Nasri M, Nezamabadi-pour H. Image denoising in the wavelet domain using a new adaptive thresholding function. Neurocomputing. 2009. January;72(4–6):1012–1025. 10.1016/j.neucom.2008.04.016 [DOI] [Google Scholar]
  • 53. Sonka M, Hlavac V, Boyle R. Image Processing, Analysis, and Machine Vision. Toronto, Canada: Thomson; 2008. [Google Scholar]
  • 54. Gottschlich C, Huckemann S. Separating the Real From the Synthetic: Minutiae Histograms as Fingerprints of Fingerprints. IET Biometrics. 2014. December;3(4):291–301. 10.1049/iet-bmt.2013.0065 [DOI] [Google Scholar]
  • 55. Gottschlich C, Schuhmacher D. The Shortlist Method for Fast Computation of the Earth Mover’s Distance and Finding Optimal Solutions to Transportation Problems. PLoS ONE. 2014. October;9(10):e110214 10.1371/journal.pone.0110214 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 56. Gottschlich C, Mihăilescu P, Munk A. Robust orientation field estimation and extrapolation using semilocal line sensors. IEEE Transactions on Information Forensics and Security. 2009. December;4(4):802–811. 10.1109/TIFS.2009.2033219 [DOI] [Google Scholar]
  • 57. Huckemann S, Hotz T, Munk A. Global models for the orientation field of fingerprints: an approach based on quadratic differentials. IEEE Transactions on Pattern Analysis and Machine Intelligence. 2008. September;30(9):1507–1517. 10.1109/TPAMI.2007.70826 [DOI] [PubMed] [Google Scholar]
  • 58.Gottschlich C, Marasco E, Yang AY, Cukic B. Fingerprint Liveness Detection Based on Histograms of Invariant Gradients. In: Proc. IJCB. Clearwater, FL, USA; 2014. p. 1–7.
  • 59. Gottschlich C, Schönlieb CB. Oriented Diffusion Filtering for Enhancing Low-quality Fingerprint Images. IET Biometrics. 2012. June;1(2):105–113. 10.1049/iet-bmt.2012.0003 [DOI] [Google Scholar]
  • 60. Fronthaler H, Kollreider K, Bigun J. Local features for enhancement and minutiae extraction in fingerprints. IEEE Transactions on Image Processing. 2008. March;17(3):354–363. 10.1109/TIP.2007.916155 [DOI] [PubMed] [Google Scholar]
  • 61. Ville DVD, Blu T, Unser M. Isotropic polyharmonic B-splines: scaling functions and wavelets. IEEE Transactions on Image Processing. 2005. November;14(11):1798–1813. 10.1109/TIP.2005.857249 [DOI] [PubMed] [Google Scholar]
  • 62. Gottschlich C. Convolution Comparison Pattern: An Efficient Local Image Descriptor for Fingerprint Liveness Detection. PLoS ONE. 2016. February;11(2):e0148552 10.1371/journal.pone.0148552 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 63.Gottschlich C, Mikaelyan A, Olsen MA, Bigun J, Busch C. Improving Fingerprint Alteration Detection. In: Proc. ISPA. Zagreb, Croatia; 2015. p. 85–88.
  • 64. Thai DH, Gottschlich C. Global Variational Method for Fingerprint Segmentation by Three-part Decomposition. IET Biometrics. [Google Scholar]
  • 65. Bartůněk JS. Fingerprint Image Enhancement, Segmentation and Minutiae Detection. Blekinge Institute of Technology; Karlskrona, Sweden; 2016. [Google Scholar]
  • 66. Thai DH, Gottschlich C. Directional Global Three-part Image Decomposition. EURASIP Journal on Image and Video Processing. 2016. March;2016(12). [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

The manually marked ground truth benchmark is available for download at http://dx.doi.org/10.6084/m9.figshare.1294209 and the Matlab implementation of the FDB method is available for download at http://dx.doi.org/10.6084/m9.figshare.1294210.


Articles from PLoS ONE are provided here courtesy of PLOS

RESOURCES