Skip to main content
The Scientific World Journal logoLink to The Scientific World Journal
. 2014 Jan 22;2014:894546. doi: 10.1155/2014/894546

Multifractal Framework Based on Blanket Method

Milorad P Paskaš 1,2,*, Irini S Reljin 1, Branimir D Reljin 2
PMCID: PMC3919058  PMID: 24578664

Abstract

This paper proposes two local multifractal measures motivated by blanket method for calculation of fractal dimension. They cover both fractal approaches familiar in image processing. The first two measures (proposed Methods 1 and 3) support model of image with embedded dimension three, while the other supports model of image embedded in space of dimension three (proposed Method 2). While the classical blanket method provides only one value for an image (fractal dimension) multifractal spectrum obtained by any of the proposed measures gives a whole range of dimensional values. This means that proposed multifractal blanket model generalizes classical (monofractal) blanket method and other versions of this monofractal approach implemented locally. Proposed measures are validated on Brodatz image database through texture classification. All proposed methods give similar classification results, while average computation time of Method 3 is substantially longer.

1. Introduction

Fractality is often introduced through geometry [1, 2] but it is valid whenever dimensionality is in the focus of research. Both geometry and calculus consider dimension as a point's relation to its neighborhood and the trend of behavior of the geometrical object or function in the surrounding of the point.

After its introduction in image processing [35] in 1980s, fractal model has been used in texture classification [6, 7], image segmentation [810], image compression [11], and edge detection [12]. Contrary to statistical models of the image, where image is perceived as a union of regions with homogeneity of chosen statistical moment (it is usually mean value of the region), fractality models an image as a union of regions with homogeneity of dimension or spectrum of dimensions. As it will be described in detail in Section 2, dimension is a nonlinear attribute of the object (region). This is of special interest when analyzing textures since the statistical moments of textural regions in an image do not describe them uniquely.

This paper proposes two local multifractal (MF) measures based on blanket method [3] for calculation of fractal dimension. Blanket method is based on monofractal model of an image, where the image is modeled as a three-dimensional surface and, hence, pixel intensity is seen as a third spatial component. The area of this surface—blanket—is assumed to be monofractal. Area is calculated by thickening the blanket and then dividing its volume by thickness of the blanket. Fractal model assumes that area of the surface is growing this way with the power of dimension. While blanket method treats an image as a monofractal object, proposed measures expand blanket method to multifractality. Each of proposed measures represents an example of two directions in implementation of fractality in image processing. Existing techniques developed within one of two approaches exclude the other approach. This paper covers mutually both approaches and employs methods for calculation of local and global dimension from both models.

The paper is organized as follows. Section 2 deals with dimensionality and introduces multifractality in general. Overview of applications of multifractal method in image processing is given in Section 3. Section 4 gives a brief survey on blanket method which inspired multifractal measures proposed in Section 5. Finally, results and experimental settings are represented in Section 6. Conclusion and further discussion on obtained results are given in Section 7. Further explanation of box-counting method and Legendre MF spectrum is given in the Appendix in order to cover all aspects of multifractality.

2. Dimensionality

2.1. Reference Model

Dimensionality is usually considered in a sense (interpretation) of geometry. Classical (nonfractal) approach considers an object's embedded dimension, that is, dimension of the hyperspace that embodies that object. Furthermore, this model implies only four discrete values of dimension: 0 for point, 1 for curve, 2 for plane, and 3 for solid body. Determination of a dimension of an object is therefore heuristic. In order to become cognizant of dimension determination task, we created simple reference model for our further discussion as shown in Figure 1. Object given in Figure 1 is a mixture of objects with different dimensions and represents a multidimensional object. This object cannot be observed as an object with only one value of dimension (unidimensional) but rather as a multidimensional set of points (or objects) with unidimensional subsets (Figure 2). This is clearer when inspecting differences between points belonging to different subsets. Thus, points from one-dimensional subsets have only one degree of freedom in sense of connectedness with other points from that subset. In other words there is only one way in connecting two distinct points within the one-dimensional object. Concept of connectedness as far as a model of dimensionality is closely related to differentiability of a function (object). It also seeks for trend of behavior [13] of a function in the neighborhood of observed point. Local trend of a function indicates topological dimension of a function or an object.

Figure 1.

Figure 1

Reference model as a setting for discussion on dimensionality. Intersections of objects of different dimensionality are permitted in this model.

Figure 2.

Figure 2

Decomposition of a reference model from Figure 1 on subsets with dimensions 0, 1, 2, and 3.

The first step toward dimensional analysis is to isolate subsets with the same dimension or singularity as explained before and then to find its dimension. The quality that discriminates subsets with different dimensions is a measure of each subset. It can be number of points within the subset, length, area, or volume. Measure is so-called set function since it maps a set of points into one point—measured value. Hence, every subset is fully described either by a measure or a dimension; these two qualities relate to the same property of an object. Significance of measure and measuring becomes apparent with objects with irregular shapes such as surfaces in Figure 1. When calculating area of any of these surfaces it must be divided into small “regular” coverings whose area is straightforwardly (in the same manner) calculated. Total area is then given through summing (integrating) all divided parts. Measure—dimension dualism can be further described as follows:

M(Fi)=FiδDi=Ni(δ)·δDi, (1a)
D(Fi)=log(M/Ni)log(δ)~log(M)log(δ), (1b)

where M(F i) is a measure of a subset F i, N i is the number of covers, and D i is the integer-valued dimension of the subset. We will recall scaling manner such as one in (1a) and (1b) many times throughout this paper. Described model of dimensionality reduces measurement on one-dimensional (unit) gauge, δ; other dimension values are deduced from the unit gauge. Thus, other order dimensions are determined in comparison to one-dimensional reference. However, this model assumes covering of a subset with identical coverings. This could be excessively strict restriction in some cases.

2.1.1. Digression

In this paper terms measure and capacity will be used interchangeably. More precisely capacity can be seen as a measure which is not additive [14]. Since mathematically correct explanation is beyond of the scope of this paper and MF formalism holds for capacities, additivity will not be examined.

2.2. Unidimensionality (Monofractality)

In mid-nineteenth century mathematicians created geometrical objects with “strange” behavior. One example is the famous Peano curve; it is a curve which fills the plane through infinite number of iterations. Its embedded dimension is surely two, although its topological dimension (dimension of the original curve) is one.

As measure and dimension inquire function integrability (i.e., differentiability) it is not possible to measure or calculate dimension for nondifferentiable points of the function. Indeed there are nondifferentiable functions in all points. An example is Weierstrass function. Therefore, dimensionality should be broadened to wider range of values. Generalization of dimensionality assumes fractional (nonintegral) value of dimension; this value is between integral values of topological and embedded dimension. Objects with non-integral dimensions are called fractals [1]. This generalization spreads on fractional derivatives but it is out of scope of this paper.

Now we can imagine set F from reference model as a union of fractal subsets (Figure 2). In order to incorporate fractional dimensions into existing model, Hausdorff [15] generalized model in (1a) and (1b):

MH(Fi,s)=limδ0infj=1Fi|Uj|s, (2a)

where {U i} is a δ-cover of unidimensional subset, F i, and s can be fractional. Dimension of a subset F i is defined correspondingly:

DH(Fi)=inf{s0:MH(Fi,s)=0}=sup{s:MH(Fi,s)=}. (2b)

This definition of dimension is stringent and it encompasses previous model given in (1a) and (1b) in case of fractional dimensions. Definition in (2b) simply says that there is only one value of s for which measure in (2a) is nontrivial. Also it implicitly shows that dimension is defined with respect to measure and it is obvious now that different measures defined on the same set will give different values of dimension.

2.3. Multidimensionality (Multifractality)

Hitherto we have been analyzing dimensions of unidimensional subsets, F i, of a multidimensional set F. It is based on assumption that we have a priori knowledge of dimensionality of each subset and we only have to calculate its quantity, that is, dimension. This solution is adequate only if the fractal synthesis is known in advance. Unfortunately, there are many situations when this assumption is lacking. Connectedness or function trend is varying on the atomic level, namely, from point to point. It leads to the conclusion that measure should be introduced in different manner than in reference model. This is the main reason for introducing local dimension or Hölder exponent of measure μ assigned to a point rather than to a set:

dμ(x)~limδ0logμ(δ(x))logδ. (3)

Local dimension is therefore a modification of global or outer dimension from (1a) and (1b) and (2a) and (2b) since it is calculated on a ball δ(x) centered in x, with radius δ, and assigned to a single point instead to a set of points. Local dimension is usually in the literature denoted with α [2]. Chosen annotation in this paper is suggestive and it points to dependence of local dimension on selected measure.

Determination of local dimension for all points within multidimensional set F categorizes points into unidimensional classes, F i, with respect to selected measure, μ:

Fi={xF:dμ(x)=di}. (4a)

Now it is possible to calculate dimension of each monofractal subset using, for example, (2a) and (2b); instead of Hausdorff measure, dimension can be calculated using any other definition of measure. Furthermore, multifractal spectrum (MFS) is defined as a relation between local and outer dimension:

DM(dμ)DM(Fi). (4b)

Dimension D is defined with respect to measure M while local dimension d is defined with respect to local measure μ. In fact, outer dimension gives information on complexity of each monofractal subset. While local dimension seeks for local singularities and trends, outer dimension is calculated on the whole set of points with the same value of the local dimension. Usually, D is Hausdorff dimension when MFS is denoted as a fine MFS, whilst d is calculated using arbitrary local measure or capacity. Multifractal spectrum is usually in the literature denoted as f(α) [2].

3. Fractal Modeling in Image Processing

There are two approaches of fractality in image processing. The first approach [3, 1618] assumes image as a surface with pixels represented as triplets (x, y, z) where x and y are spatial coordinates and z is pixel intensity. Thus, pixel intensity is treated as two spatial coordinates. In turn this involves embedded dimension to be 3 and treats an image geometrically. Second approach [14] models pixels as triplets (x, y, f(x, y)) where x and y are again spatial coordinates while f(x, y) is pixel intensity and it is modeled as a function over spatial coordinates. An image is seen as an object embedded in 2D space with strict distinction between spatial and pixel intensity components. This model is closer to traditional image processing techniques.

Both approaches address model described in the previous section: in the first approach the dimension is calculated for the 3D object from the geometrical viewpoint via area measuring and calculations are done in 3D space, while in the second case the dimension is calculated for the function f(x, y) that models pixel intensities in the 2D space and μ H(f(F i), s) is calculated.

4. Blanket Method

Here we shall give a brief overview of blanket method proposed for calculating dimension of a monofractal set of points [3]. Blanket method represents an example of the first (3D embedded space) approach from previous section. The main idea of this approach of calculating fractal dimension of unidimensional set of points is to employ area of the image as a global measure. Area is computed indirectly via volume, that is, blanket constructed around original image surface through successive iterations of thickening of the blanket. Upper, u ε(x, y), and lower surface, b ε(x, y), of the blanket in εth iteration, respectively, are given by

uε(x,y)=max{uε1(x,y)+1,maxS1(x,y)uε1(x,y)}, (5a)
bε(x,y)=min{bε1(x,y)1,minS1(x,y)bε1(x,y)}, (5b)
u0(x,y)=b0(x,y)=f(x,y). (5c)

In these equations S 1(x, y) represents 3 × 3 pixel neighborhood and f(x, y) denotes pixel intensity. The first term in the brackets (addition or subtraction of 1) ensures thickening of the blanket in every iteration. Volume of the blanket is calculated as a sum of differences:

Vε(Fε)=Fε[uε(x,y)bε(x,y)]. (5d)

Finally, measure (area) is given by dividing volume by minimal thickness of the blanket, ε:

MBLANKET(Fε)=Vε2ε. (5e)

Although the gauge in this method is of dimension 3 (blanket is three dimensional object), the final measure is of dimension 2, since the area is calculated ultimately. If we recall here the MF principles, then thickness of the blanket corresponds to scale and object is regarded as a monofractal a priori.

5. Multifractal Blanket Method

5.1. Previous Work

Blanket method is widely used in image processing through many different approaches. In [19, 20] image is divided in blocks (subimages) and FD is calculated for each block by direct implementation of blanket method. Local fractal dimension (LFD) for blanket method is proposed in [21]. Here, locality considers the gliding window of fixed size on which the classical blanket method is implemented for each pixel, that is, only blanket thickness is scaling. Obtained matrices are called LFD maps. LFD concept is generalized for volumetric textures in [22]. Further multifractal (global fractal) analysis for LFD is not provided in these papers.

LFD notion implies pixel-wise implementation of so-called global fractal dimension, that is, classical blanket method described in previous section. This concept is quasimultifractal approach.

Concerning multifractal approach, it is presented in the literature mainly through use of coarse multifractal spectrum [23], especially Legendre MFS (see the Appendix). Calculation of fractal dimension is mainly based on box-counting (BC) capacity and its modifications, for example, differential box-counting (DBC) capacity. All these capacities assume an image embedded in 3D space. Fine MFS is used only when image is modeled as a 2D object [9, 24]. In this section, new local multifractal measures of both types are proposed and are inspired by classical blanket method.

5.2. Proposed 3D Multifractal Framework (Method 1)

Instead of a pixel-wise implementation of blanket method, here is proposed a local measure based on blanket method which can be used for calculation of local dimension in a sense of fine MF formalism described in (3).

In classical blanket method upper and lower surfaces (5a), (5b), alternatively, can be rewritten as follows:

uε(x,y)=max{u0(x,y)+ε,maxSε(x,y)u0(x,y)}, (6a)
bε(x,y)=min{u0(x,y)ε,minSε(x,y)u0(x,y)}. (6b)

This equation gives a possibility of defining a local measure inspired by blanket method. Hence, the local volume is computed over an ε—neighborhood (ε × ε pixels neighborhood) of each pixel, S ε(x, y), and it represents local blanket measure:

μBLANKET(ε(x))=Sε(x,y)[uε(x,y)bε(x,y)]. (6c)

It is worth saying that not only the thickness of the blanket is scaled but also simultaneously the pixel neighborhood is scaled according to MF formalism.

5.3. Proposed 2D Multifractal Framework (Method 2)

Measure presented in (6a), (6b), and (6c) assumes an image as a surface embedded in 3D space, since measure considers intensities interchangeably with spatial coordinates. In order to create a measure for model of an image where the pixel intensities are seen as a function over spatial coordinates, equations in (6a) and (6b) are revisited as

uε(x,y)=max{w1·u0(x,y),maxSε(x,y)u0(x,y)}, (7a)
bε(x,y)=min{w2·u0(x,y),minSε(x,y)u0(x,y)}, (7b)

where w 1 and w 2 are weights which simulate the expansion of the blanket in (6a), (6b), and (6c). Equations in (7a) and (7b) are not depending on blanket thickness. Local dimension can be computed using (6c) and the fine MFS as defined in (4a) and (4b).

Further inspection of this measure indicates its correlation with three well known capacities: MIN, MAX, and SUM [9]. It is indeed superposition of these three capacities with correction term causing monotony of the measure.

5.4. Blanket Multifractal Spectrum (Method 3)

Classical blanket method considers an image as a monofractal object. Otherwise if multifractality is taken into consideration, there should be used decomposition of the image on monofractal partitions. Local measure can be chosen arbitrary. For each of the partitions blanket measure (5a), (5b), (5c), (5d), and (5e) and blanket dimension are then calculated. Obtained spectrum can be seen as the blanket multifractal spectrum. The only restriction is to use local measure based on approach where pixel intensity is modelled as a third spatial coordinate.

6. Experimental Results

Mathematical model of multifractality given in (4a) and (4b) assumes infinitesimal range of scales. There should be made a difference between original data (infinitesimally increasing resolution of a view of a scene) and its sampled version (digital image) [5]. Image resolution is finite and some numerical assumptions should be considered. Therefore, windows and box sizes (or scales) used in calculation of local and global dimension, respectively, of the image are discussed in this section in detail. Performed analysis is focused primarily on discrimination potential of proposed MF spectra.

This section ends with examination of appropriateness of proposed methods for calculation of MF spectra to discriminate different textures. In applications dimension and MFS are used as signatures of images and implemented for discrimination. Fractal modelling in texture classification is commonly implemented through calculation of fractal dimension (FD), where each texture is described by single value (FD value). Since two different textures can be of the same dimensionality, one more value—lacunarity [1, 25, 26]—is usually employed. Some authors [1, 27] propose use of succolarity for better discrimination. Multifractal spectrum is often used as a discriminator [9, 10].

6.1. Least Squares Approximation

Calculation of dimension, regardless of local or global dimension, is based on power law. This is clear from (1a) and (1b), (2a) and (2b), and (3) and it can be in general written as

measure=C·scaledimension (8a)

or after taking logarithm on both sides as

log(measure)=log(C)+dimension·log(scale). (8b)

Calculation of dimension in this model implies determining measure variations while varying scale. Our model from Section 2 assumes dimension and C to be constants. In order to facilitate writings, substitution of variables provides

y^=k·x+n, (8c)

where y^=log(measure), k = dimension, x = log⁡⁡(scale), and n = log⁡⁡(scale). Therefore, our model in (8c) predicts values yi^ while measured values (of logarithm of measure) are y i. Determination of value k (i.e., dimension) in fractal modeling is usually done through minimisation of least squares error [28] defined as

ei=1N{yiyi^}2, (9)

where N is the total number of scales. Differentiating least squares error both for k and n and putting derivatives to zero (∂e/∂k = 0, ∂e/∂n = 0) gives a dimension value

k=Ni=1Nxiyi{i=1Nxi}{i=1Nyi}Ni=1Nxi2{i=1Nxi}2. (10)

6.2. Multifractal Spectrum Calculation

In this section all MF spectra are calculated using histogram method [2, 28, 29] and box-counting (BC) capacity (see the Appendix). Actually, here numerical evaluation of equations given in (4a) and (4b) is shortly described. After determining local dimension values, d i, using least squares approximation, there is a range of local dimension values, d ∈ (d min⁡, d max⁡). According to (4a) and (4b) regions in image with the same value of local dimension should be exploited for calculation of global dimension. Straightforward application of that principle is unacceptable and some rounding must be done. The easiest solution is to uniformly quantize local dimension range to arbitrary resolution (number of uniformly distant levels), R. For each region with the same local dimension value, global fractal dimension is then calculated.

6.3. Multifractal Spectra Inspection

Local dimension considers local behaviour of the points (pixels) and consequently used rectangular windows are supposed to be lower. Behaviour of the BC MF spectrum with respect to different windows sizes in calculating local dimension is shown in Figures 3 and 4, for Method 1 and Method 2, respectively. Image analysed in this section is of size 640 × 640 pixels. Parameters for calculating MFS are the same for each curve, while only scales for local dimension calculation are varying. From Figures 3 and 4 it is evident MFS shrinkage while the form of the MFS does not change significantly.

Figure 3.

Figure 3

BC MFS obtained using Method 1 for different window sizes, w, for calculation of local dimension. Notation A : B implies set of values A, A + 1,…, B.

Figure 4.

Figure 4

BC MFS obtained using Method 2 for different windows sizes, w, for calculation of local dimension. Notation A : B implies set of values A, A + 1,…, B.

Weights w 1 and w 2 in proposed Method 2 and their influence on BC MFS are analyzed in Figures 5 and 6. Parameter w 1 is always greater than 1 while parameter w 2 is smaller than 1. The reason for this is in providing monotony as explained in Section 5. Hence, parameter w 1 avoids upper blanket to be lower in every successive iteration. The same stands for parameter w 2 and lower blanket.

Figure 5.

Figure 5

BC MFS obtained using Method 2 for different values of weight w 1.

Figure 6.

Figure 6

BC MFS obtained using Method 2 for different values of weight w 2.

The influence of chosen box sizes on calculation of BC MF spectra in Method 1 and Method 2 is shown in Figures 7 and 8. Higher ranges of window sizes produce spectra which are not discriminative regarding wide range of local dimension values. On the other hand small windows disregard global behaviour of the image and accent only local singularities.

Figure 7.

Figure 7

BC MFS obtained using Method 1 for different box sizes, w, for calculation of global dimension. Notation A : B : C implies set of values A, A + B,…, C.

Figure 8.

Figure 8

BC MFS obtained using Method 2 for different box sizes, w, for calculation of global dimension. Notation A : B : C implies set of values A, A + B,…, C.

Finally, MF blanket spectrum is analyzed for different blankets (Figure 9). After blanket of size 30, higher values of blanket do not induce significant changes in the shape of MFS and there is evident rounding of MFS curve.

Figure 9.

Figure 9

Blanket MFS (Method 3) obtained for different blankets, e.

6.4. Texture Discrimination

Each dimension and measure describes particular singularity and this section explores potentials of proposed measures. As an illustration of MF spectra diversity, spectra of three random textures from [30] are calculated and shown in Figures 10, 11, and 12, for all three proposed methods. When calculating MF according to Method 1 and Method 2, BC method is used for calculation of global dimension. Method 3 assumes local dimension calculation from Method 1.

Figure 10.

Figure 10

BC MFS obtained with Method 1 for three different textures from [30].

Figure 11.

Figure 11

BC MFS obtained with Method 2 for three different textures from [30].

Figure 12.

Figure 12

Blanket MFS for three different textures from [30]. Local dimension values are calculated using local MF measure from Method 1.

Proposed three methods are tested on Brodatz database [30, 31] with 111 texture images. Each texture (640 × 640 pixels) is divided into 25 images (128 × 128 pixels). Within obtained database of totally 2775 images two different textures (25 images of one and 25 of the other texture) are classified. For classification K-means method [32] is used. Since this method highly depends on initial cluster centroid positions, classification is repeated for 400 times. As a result of classification the solution is used with the lowest sum of distances to every centroid. Correctness rate is calculated as a ratio between correctly classified images within both textures and the total number of images (50). Results of classification for all pairs of textures in the database are illustrated in Figures 13, 14, and 15. In matrices in these figures color of each point corresponds to the percent of correctly classified images belonging to one of two Brodatz textures defined by abscissa and ordinate values. Table 1 brings mean correctness rate for all three methods with variation of some parameters in calculation of MFS. For Methods 1 and 2 both sizes of boxes (MFS scales) and resolution of MF spectrum vary. Average computation time of features in Method 3 is the main reason why only one test is done. As it is obvious from Table 1 obtained precision is even inferior to the other two methods. Computations are done on Intel Core i5 3 GHz processor.

Figure 13.

Figure 13

Percentage of correctly classified pairs of textures from the Brodatz database for Method 1. Texture order corresponds to [31].

Figure 14.

Figure 14

Percentage of correctly classified pairs of textures from the Brodatz database for Method 2. Texture order corresponds to [31].

Figure 15.

Figure 15

Percentage of correctly classified pairs of textures from the Brodatz database for Method 3. Texture order corresponds to [31].

Table 1.

Comparison of proposed three methods regarding precision and computation time.

Method MFS scales MFS resolution Mean correctness (%) Average time per pixel (d values) (µs/pix) Average time per pixel (D values) (µs/pix)
1 2 : 10 : 72 20 92.2234 1.8443
20 93.7808 15.5874 1.9072
2 : 2 : 20 15 93.1751 1.4749
10 93.5119 1.0510

2 2 : 10 : 72 20 90.9808 3.2328
20 92.4177 15.2868 3.4039
2 : 2 : 20 15 92.1202 2.6124
10 91.5135 1.8256

3 7 20 88.4986 15.5874 39.0058

All experiments are performed on processor Intel Core i5, 3 GHz.

Notation in second column A : B : C indicates values A, A + B, A + 2B,…, C.

All d values were calculated using window sizes 2, 3, 4, and 5. In Method 2 parameters w 1 and w 2 are equal to one.

Comparing results of classification between different methods (Table 1), Methods 1 and 2 are superior regarding both correctness rate and average computation time. Classification is performed also with classical blanket method [3] and obtained mean correctness is 84.5281% (maximum blanket thickness is 20). From Figures 13, 14, and 15 it is evident that misclassifications occur almost for the same pairs of textures in all three scenarios. One of the reasons for lower classification rates is high variation between images within the same class. If the texture pattern is large then images obtained by division of this texture will substantially differ.

7. Conclusion

Proposed multifractal measures correspond to two models in fractal modeling of images. To our knowledge measure introduced within Method 1 is the first local measure assuming pixel intensity as a third spatial coordinate whose fine MFS is calculated. Usually only Legendre MFS is calculated for such measures. Rearranging this measure into measure with pixel intensity modeled as a function over spatial coordinates transposes Method 1 to Method 2 through dimension reduction. Conducted tests suggest nearly identical results according to both precision and time consumption. Further reduction of computation time could be done by calculation of MFS only for a region of a texture, since there is a repetition of texture's pattern within the texture. As results show this must be done empirically for each texture because of the ratio between sizes of the pattern and the whole texture.

Acknowledgment

This work was partially supported by the Ministry of Education, Science and Technological Development, Government of the Republic of Serbia, under the Project III 44009.

Appendix

This appendix brings simple and efficient technique for estimating Hausdorff MFS via Legendre transform with aim to comprehend MFS calculation context. In the literature it is usually but mistakenly introduced as MF formalism already described by (3) and (4a) and (4b). Legendre MFS has its origin in statistical mechanics and terminology used in derivation of this spectrum is taken from there.

Hausdorff measure is given by

MH(Fi,s)=limδ0infj=1Fi|Uj|s. (A.1)

Since δ-covering is hardly realizable in practice, when there is no knowledge of fractal object synthesis (creation), it is reasonable to tile the object with covers of the same length, δ:

MB(Fi,s)=limδ0infFiδs. (A.2)

M B is called box-counting capacity and dimension resulting from it is called box-counting dimension, D B. Box-counting dimension is an overestimation of Hausdorff dimension [2]:

DH(Fi)DB(Fi). (A.3)

Box-counting capacity leads to so-called coarse MFS. It is often desirable to calculate MFS in specific points such as a maximum of MFS function or in a point with D = d. Shapes of MF spectra of different objects (images) can be compared in space of particular points. With this motivation let us transform function D B of variable d μ into another variable, which represents first derivative of dimension, q = dD B/dd μ. Legendre transform provides it:

(DB(dμ))=DB(q)=infdμ{q·dμDB(dμ)}. (A.4)

Alternatively Legendre transform can be defined using supremum (maximum) instead of infimum (minimum) [33].

Transformation given with (D B, D B*) is Legendre transformation pair, hence, the name of the corresponding spectrum. Figure 16 shows how Legendre transform centers maximum of spectrum into zero in space of q-variable. If D B is concave, that is, ∂2 D B/∂d μ 2 ≤ 0, then (A.4) is valid without brackets. Legendre transform is involutive [34], that is, Legendre transform of D B* is D B. In other words direct and inverse Legendre transforms are identical. Substitution of transformed dimension gives the following capacity:

ML(Fi,DB)=δDBFiδq·dμ=δDBFiμ(δ(x))q,ML(Fi,DB)δDBFμ(δ(x))q. (A.5)

The term on the right side of inequality, that is, sum of moments, stems for partition function [35], Γ(q) = ∑F μ( δ(x))q, from the analogy with thermodynamics. It depends only on q but not on scale, δ. Thus, Legendre transform of dimension is

DB(q)~logFμ(δ(x))qlogδ. (A.6)

This variable is often labeled in literature as −β(q) [2] or τ(q) [35]. Applying chosen measure μ on blocks within the mesh grid of resolution δ and calculating qth power moment sums of μ gives the value of Legendre transform for each q. Legendre transform of D B*(q) gives MFS, D B. Local dimension values, needed for calculation of D B, are also determined via D B* by equating ∂D B/∂q = 0:

dμ=DBq, (A.7)

therefore,

DB=q·DBqDB. (A.8)

As it is obvious from the derivation of Legendre MFS, this numerical method provides only pairs in the graph D B(d μ), without calculating local and global dimension values for each point of the signal (function). This method executes faster on computer although the precision is sacrificed.

Figure 16.

Figure 16

Legendre transform illustration. Red and blue domains evoke regions with positive or negative values for q. Here Legendre transform curve is a reflected image of original function, but it is not true in general.

Legendre MFS can be calculated not only by box-counting method but also by using gliding-box method [36]. This method calculates partition function on slipping box, that is, for every pixel. Reported results are better than using box-counting calculation.

Conflict of Interests

The authors declare that there is no conflict of interests regarding the publication of this paper.

References

  • 1.Mandelbrot BB. The Fractal Geometry of Nature. New York, NY, USA: W. H. Freeman; 1983. [Google Scholar]
  • 2.Falconer K. Fractal Geometry: Mathematical Foundations and Applications. 2nd edition. John Wiley and Sons; 2003. [Google Scholar]
  • 3.Peleg S, Naor J, Hartley R, Avnir D. Multiple resolution texture analysis and classification. IEEE Transactions on Pattern Analysis and Machine Intelligence. 1984;6(4):518–523. doi: 10.1109/tpami.1984.4767557. [DOI] [PubMed] [Google Scholar]
  • 4.Pentland AP. Fractal-based description of natural scenes. IEEE Transactions on Pattern Analysis and Machine Intelligence. 1984;6(6):661–674. doi: 10.1109/tpami.1984.4767591. [DOI] [PubMed] [Google Scholar]
  • 5.Keller JM, Crownover RM, Chen RY. Characteristics of natural scenes related to the fractal dimension. IEEE Transactions on Pattern Analysis and Machine Intelligence. 1987;9(5):621–627. doi: 10.1109/tpami.1987.4767956. [DOI] [PubMed] [Google Scholar]
  • 6.Wu C-M, Chen Y-C, Hsieh K-S. Texture features for classification of ultrasonic liver images. IEEE Transactions on Medical Imaging. 1992;11(2):141–152. doi: 10.1109/42.141636. [DOI] [PubMed] [Google Scholar]
  • 7.Kaplan LM. Extended fractal analysis for texture classification and segmentation. IEEE Transactions on Image Processing. 1999;8(11):1572–1585. doi: 10.1109/83.799885. [DOI] [PubMed] [Google Scholar]
  • 8.Chen C-C, Daponte JS, Fox MD. Fractal feature analysis and classification in medical imaging. IEEE Transactions on Medical Imaging. 1989;8(2):133–142. doi: 10.1109/42.24861. [DOI] [PubMed] [Google Scholar]
  • 9.Vehel JL, Mignot P. Multifractal segmentation of images. Fractals. 1994;2(3):371–377. [Google Scholar]
  • 10.Stojić T, Reljin I, Reljin B. Adaptation of multifractal analysis to segmentation of microcalcifications in digital mammograms. Physica A. 2006;367:494–508. [Google Scholar]
  • 11.Thomas L, Deravi F. Region-based fractal image compression using heuristic search. IEEE Transactions on Image Processing. 1995;4(6):832–838. doi: 10.1109/83.388086. [DOI] [PubMed] [Google Scholar]
  • 12.Turiel A, Parga N. The multifractal structure of contrast changes in natural images: from sharp edges to textures. Neural Computation. 2000;12(4):763–793. doi: 10.1162/089976600300015583. [DOI] [PubMed] [Google Scholar]
  • 13.Struzik ZR. Wavelet methods in (financial) time-series processing. Physica A. 2001;296(1-2):307–319. [Google Scholar]
  • 14.Vehel JL. Fractal approaches in signal processing. Fractals. 1995;3(4):755–775. [Google Scholar]
  • 15.Hausdorff F. Dimension and outer measure. In: Edgar GA, editor. Classics on Fractals. Westview Press; 2004. pp. 75–99. [Google Scholar]
  • 16.Clarke KC. Computation of the fractal dimension of topographic surfaces using the triangular prism surface area method. Computers and Geosciences. 1986;12(5):713–722. [Google Scholar]
  • 17.Dubuc B, Zucker SW, Tricot C, Quiniou JF, Wehbi D. Evaluating the fractal dimension of surfaces. Proceedings of the Royal Society A. 1989;425:113–127. [Google Scholar]
  • 18.Sarkar N, Chaudhuri BB. Multifractal and generalized dimensions of gray-tone digital images. Signal Processing. 1995;42(2):181–190. [Google Scholar]
  • 19.Zheng L, Chan AK. An artificial intelligent algorithm for tumor detection in screening mammogram. IEEE Transactions on Medical Imaging. 2001;20(7):559–567. doi: 10.1109/42.932741. [DOI] [PubMed] [Google Scholar]
  • 20.Zook JM, Iftekharuddin KM. Statistical analysis of fractal-based brain tumor detection algorithms. Magnetic Resonance Imaging. 2005;23(5):671–678. doi: 10.1016/j.mri.2005.04.002. [DOI] [PubMed] [Google Scholar]
  • 21.Novianto S, Suzuki Y, Maeda J. Near optimum estimation of local fractal dimension for image segmentation. Pattern Recognition Letters. 2003;24(1–3):365–374. [Google Scholar]
  • 22.Lopes R, Dubois P, Bhouri I, Bedoui MH, Maouche S, Betrouni N. Local fractal and multifractal features for volumic texture characterization. Pattern Recognition. 2011;44(8):1690–1697. [Google Scholar]
  • 23.Xia Y, Feng D, Zhao R. Morphology-based multifractal estimation for texture segmentation. IEEE Transactions on Image Processing. 2006;15(3):614–623. doi: 10.1109/tip.2005.863029. [DOI] [PubMed] [Google Scholar]
  • 24.Xu Y, Fermüller C. Viewpoint invariant texture description using fractal analysis. International Journal of Computer Vision. 2009;83(1):85–100. [Google Scholar]
  • 25.Vehel JL. About lacunarity, some links between fractal and integral geometry, and an application to texture segmentation. Research Report. 1990;(INRIA)
  • 26.Kilic KI, Abiyev RH. Exploiting the synergy between fractal dimension and lacunarity for improved texture recognition. Signal Processing. 2011;91(10):2332–2344. [Google Scholar]
  • 27.Melo RHC, Conci A. How succolarity could be used as another fractal measure in image analysis. Telecommunication Systems. 2013;52(3):1643–1655. [Google Scholar]
  • 28.Turner MJ, Blackledge JM, Andrews PR. Fractal Geometry in Digital Imaging. Academic Press; 1998. [Google Scholar]
  • 29.Vehel JL. Numerical computation of the large deviation multifractal spectrum. Proceedings of the Conference on Future Internet Communications (CFIC '96); 1996; Rome, Italy. [Google Scholar]
  • 30.Brodatz P. Texture: A Photographic Album for Artists and Designers. New York, NY, USA: Dover; 1966. [Google Scholar]
  • 31.Brodatz textures. http://www.ux.uis.no/~tranden.
  • 32.Seber GAF. Multivariate Observations. New York, NY, USA: John Wiley and Sons; 1984. [Google Scholar]
  • 33.Touchette H. Research Report. New York, NY, USA: Rockefeller University; 2007. Legendre-Fenchel transforms in a nutshell. http://www.maths.qmul.ac.uk/~ht/archive/lfth2.pdf. [Google Scholar]
  • 34.Arnold VI. Mathematical Methods of Classical Mechanics. 2nd edition. Springer; 1989. [Google Scholar]
  • 35.Halsey TC, Jensen MH, Kadanoff LP, Procaccia I, Shraiman BI. Fractal measures and their singularities: the characterization of strange sets. Physical Review A. 1986;33(2):1141–1151. doi: 10.1103/physreva.33.1141. [DOI] [PubMed] [Google Scholar]
  • 36.Cheng Q. The gliding box method for multifractal modeling. Computers and Geosciences. 1999;25(9):1073–1079. [Google Scholar]

Articles from The Scientific World Journal are provided here courtesy of Wiley

RESOURCES