Skip to main content
Computational and Mathematical Methods in Medicine logoLink to Computational and Mathematical Methods in Medicine
. 2019 Dec 3;2019:4025305. doi: 10.1155/2019/4025305

Biases in the Simulation and Analysis of Fractal Processes

Clément Roume 1, Samar Ezzina 1,2, Hubert Blain 1,3, Didier Delignières 1,
PMCID: PMC6914972  PMID: 31885679

Abstract

Fractal processes have recently received a growing interest, especially in the domain of rehabilitation. More precisely, the evolution of fractality with aging and disease, suggesting a loss of complexity, has inspired a number of studies that tried, for example, to entrain patients with fractal rhythms. This kind of study requires relevant methods for generating fractal signals and for assessing the fractality of the series produced by participants. In the present work, we engaged a cross validation of three methods of generation and three methods of analysis. We generated exact fractal series with the Davies–Harte (DH) algorithm, the spectral synthesis method (SSM), and the ARFIMA simulation method. The series were analyzed by detrended fluctuation analysis (DFA), power spectral density (PSD) method, and ARFIMA modeling. Results show that some methods of generation present systematic biases: DH presented a strong bias toward white noise in fBm series close to the 1/f boundary and SSM produced series with a larger variability around the expected exponent, as compared with other methods. In contrast, ARFIMA simulations provided quite accurate series, without major bias. Concerning the methods of analysis, DFA tended to systematically underestimate fBm series. In contrast, PSD yielded overestimates for fBm series. With DFA, the variability of estimates tended to increase for fGn series as they approached the 1/f boundary and reached unacceptable levels for fBm series. The highest levels of variability were produced by PSD. Finally, ARFIMA methods generated the best series and provided the most accurate and less variable estimates.

1. Introduction

The repeated measurement of physiological or behavioral events (stride durations, heartbeat intervals, and intertap intervals) is typically characterized by a marked variability. For a long time, this variability has just been considered a random perturbation, without any functional signification. However, a number of authors, during the last two decades, showed that these biological series presented a typical long-range correlational structure over time and especially a statistical self-similar (fractal) pattern [14]. Fractal processes have recently received a growing interest, especially in the domain of rehabilitation. More precisely, the evolution of fractality with aging and disease, suggesting a loss of complexity [5], has inspired a number of studies that tried, for example, to entrain patients with fractal rhythms [68].

In this domain, authors are confronted with two main methodological problems: The first one concerns the evaluation of the level of long-range correlations in physiological series. A number of different methods have been proposed, and their respective qualities were systematically assessed in comparative studies [912]. The second one refers to the generation of exact fractal signals necessary of providing experimental devices (metronomes and virtual environments) with controlled long-range correlation properties. A number of methods have been proposed for simulating such series [1315]. The assessment of estimation methods, on the one hand, and simulation methods, on the other hand, raises a typical problem of circularity, performances, and biases in the former being analyzed on the basis of the latter, and vice versa [11]. When a bias is identified, it remains difficult to attribute the problem to the method of simulation or to the method of assessment. In order to overcome this problem, we propose in the present paper a cross-validation study, combining three methods of generation and three methods of analysis.

We first propose a formal introduction of the three main domains of definition of long-range correlated processes: the fractional Brownian motion framework [16], the spectral domain [17], and the autoregressive fractionally integrated moving average (ARFIMA) processes [18].

2. Theoretical Models

2.1. The fBm/fGn Model

The fractional Brownian motion (fBm) denoted as BH(t) is a mathematical model of continuous stochastic process introduced by Mandelbrot and Van Ness [16] as a generalization of the Brownian motion, where increments do not need to be independent. BH(t) is characterized by the Hurst parameter (H), which can take any real value within the interval ]0, 1[. This value gives information about the nature and the strength of the correlation between successive increments in the process. If H is below 0.5, BH(t) is underdiffusive, and its increments are anticorrelated. In contrast, if H is above 0.5, BH(t) is overdiffusive, and its increments are positively correlated. In the case that H equals 0.5, B0.5(t) is an ordinary Brownian motion (normal diffusion), and its increments are an uncorrelated Gaussian white noise.

fBm has some fundamental properties. The first one is that its variance grows as a power function of the length of the time interval observed, with an exponent 2H:

VBHΔt2H. (1)

The second is that fBm is a fractal process, characterized by statistical self-similarity:

BHataHBHt,a  and  t>0. (2)

By definition, BH(t) is fully described by its autocovariance function γBH(t, s):

γBHt,s=covBHt,BHs=VBH2t2H+s2Hts2H. (3)

As previously indicated, the successive increments of BH(t) can be correlated and the Hurst exponent informs about the nature of this correlation, and thus, the derivative of BH(t) should be a stationary correlated noise. However, this derivative cannot be calculated because in theory Brownian motion describes an infinitely broken continuous trajectory. In other words, nondifferentiability is a fundamental property of BH(t). One can still estimate this derivative using a discrete version of BH(t) where increments are defined on a time interval m. This estimate is a discrete process called fractional Gaussian noise (fGn), denoted as Gm(i):

Gmi=BHtBHtm. (4)

As for the fBm, the fGn is fully described by its autocovariance function, which is easily derived from equation (3) with m = 1:

γG1τ=covG1i;G1i+τ=VG12τ+12H2τ2H+τ12H. (5)

2.2. The Spectral Model

Stochastic fractal processes can also be defined in the frequency domain, on the basis of a scaling law that relates power (i.e., squared amplitude) to frequency according to an inverse power function, with an exponent β [9, 10]. For an fGn process with exponent H, the power spectrum can be expressed as follows [12, 19]:

Sf1f2H1, (6)

and for the corresponding fBm process [12, 20]:

Sf1f2H+1. (7)

Then the power spectrum of fGn/fBm processes has the general form:

Sf1fβ, (8)

with β ∈ ]−1, 1[ for fGn and β ∈ ]1, 3[ for fBm. This suggests that fGn and fBm could be considered as a continuum, with both families being characterized in the time domain by a scaling exponent α, α = H for fGn and α = H + 1 for fBm. This assumption has been exploited by Peng et al. [4] in the conception of the detrended fluctuation analysis (see Section 3). The scaling exponent α is linearly related to the spectral exponent β over the whole fGn/fBm continuum by α=(β + 1/2) [21]. The case β = 1 defines the so-called “1/f noise,” which represents the boundary between fGn and fBm processes.

2.3. The ARFIMA Model

A third approach to long-range correlated processes is provided by the autoregressive fractionally integrated moving average (ARFIMA) models [18]. This approach is an extension of the ARIMA (for autoregressive, integrated, moving average) framework, introduced by Box et al. [22], which intended to represent a variety of short-term relationships in time series. ARIMA models are potentially composed of three components. The autoregressive component suggests that the current observation y(t) is determined by a weighted sum of the p previous observations, plus a random perturbation ε:

yt=k=1pϕkytk+εt. (9)

The moving average component supposes that the current observation depends on the value of the random perturbations that affect the q preceding observations, plus its own specific perturbation:

yt=k=1qθkεtk+εt. (10)

Finally, the differencing parameter d indicates the number of differencing that should be applied to the series before modeling. An ARIMA model is the combination of these three components and can be designated by the respective orders of the three processes as (p, d, q). ARIMA modeling has been used either for generating time series with specified p, d, q parameters or for determining the best p, d, q combination for accounting for a given series [22, 23].

ARIMA models could be more conveniently expressed using the so-called backshift operator, defined as

Bxt=xt1. (11)

The generic ARIMA (p, d, q) model can then be rewritten as

ϕB1Bdxt=θBεt, (12)

where ϕ(B) and θ(B) are, respectively, the autoregressive and the moving average operators, represented as polynomials in the backshift operator: ϕ(B)=1 − (1) − B2ϕ(2) − ⋯− Bpϕ(p) and θ(B)=1+(1)+ B2θ(2)+⋯+Bqθ(q) [24]. In the initial formulation of the model, the d parameter was an integer [22]. Granger and Joyeux [18] showed that it was possible to provide this model with long-range dependence properties by allowing the differencing parameter d to take on fractional values, thereby obtaining an ARFIMA model.

Here we focus on the most simple model ARFIMA(0, d, 0), which is supposed to only contain long-range correlations. Using the backshift operator notation, this model is expressed as follows [25]:

1Bdxt=εt, (13)

with

1Bd=k=0dk1kBk. (14)

Granger and Joyeux [18] derive a filter A(B) from equations (13) and (14) and demonstrate that the process can be rewritten as

xt=ABεt=k=0akεtk, (15)
ak=Γk+dΓk+1Γd. (16)

where d is a measure of the intensity of long-range correlations in the series. Note, however, that ARFIMA models account only for stationary series, d being bounded within the interval ]−0.5; 0.5[. In other words, ARFIMA models remain limited to fGn series. d is related to the spectral exponent β and the scaling exponent α by the following linear equations:

β=2d,α=2d+12. (17)

Each of these theoretical frameworks provided specific methods for generating fractal signals or for assessing the fractality of empirical series. In the present work, we engaged a cross validation of three methods of generation and three methods of analysis. We selected one simulation method and one estimation method in each previously presented domain of definition. Our rational is that biases that are revealed by the three analysis methods should be attributed to the generation method and conversely biases that appear whatever the generation method should originate in the analysis method.

3. Methods

In order to explore the whole fGn/fBm continuum, we first generated series from α = 0.1 to 0.9, by steps of 0.1, and from 1.01 to 1.9, by steps of 0.1. These values were used in most previous similar studies [9, 10, 26]. Additionally, in order to analyze more closely the behavior of simulation and analysis methods around to the 1/f boundary, we generated a series from α = 0.91 to 1.09, by steps of 0.01. This range of exponents was rarely considered in the literature (for a noticeable exception, see [27]). However, this focus on the 1/f boundary seems of particular interest, because the problems of fGn/fBm classification are concentrated within this interval [10] and also because one could have some doubts about the hypothesis of continuity around the 1/f boundary [28].

We used three methods of generation: the Davies–Harte algorithm, the spectral synthesis method, and the ARFIMA simulation method. These methods are detailed below. For each selected α value and with each method, we generated 120 series of 1024 data points. In this section, all methods are written in the discrete time and frequency domain; for reading convenience, we keep the variable t for discrete time domain and f for discrete frequency domain.

3.1. Davies–Harte Algorithm (DH)

We used the algorithm proposed by Davies and Harte [13], for generating fGn series of length N (N being a power of 2). As previously indicated, an fGn process is fully described by its autocovariance function (see equation (5)). Then, one can deduce the exact spectral power S expected for this autocovariance function, from the discrete Fourier transform of the following sequences of covariance values γG1 defined by equation (5): γG1(0), γG1(1),…, γG1((N/2)−1); γG1(N/2), γG1((N/2)−1),…, γG1(1):

Sf=τ=0N/21γG1τei2πfτ/N+τ=N/2N1γG1Nτei2πfτ/N, (18)

where f = 0, 1,…, N − 1. It is important to check that S(f) ≥ 0 for all f. Negativity would indicate that the sequence is not valid.

Let Wgn(f), where f = 0,…, N − 1, be a white Gaussian noise. The randomized spectral amplitudes, V(f), are calculated according to the following equations:

V0=NS0×Wgn0,Vf=NSf2×Wgn2f1+iWgn2f for 1fN2,VN2=NSN2×WgnN1,Vf=NSf2×Wgn2N12fiWgn2N2f for N2+1fN1. (19)

Finally, the first N elements of the discrete Fourier transform of V are used to compute the simulated series x(t):

xt=1Vf=1Nf=0N1Vfei2πft/N, (20)

where t = 1, 2,…, N.

We first generated fGn series for H values ranging from 0.1 to 0.9, by steps of 0.1. In order to explore more precisely the performance of DFA close to the 1/f boundary, we also generated fGn series for H values ranging from 0.91 to 0.99, by steps of 0.01. A second set of fGn series was generated, for H values ranging from 0.1 to 0.9, by steps of 0.1, and was integrated for obtaining fBm series for each corresponding H value (i.e., for α values ranging from 1.1 to 1.9). Finally, we generated fGn series for H ranging from 0.01 to 0.09 and integrated them for obtaining fBm series close to the 1/f boundary (i.e., for α values ranging from 1.01 to 1.09).

3.2. Spectral Synthesis Method (SSM)

The spectral synthesis method is designed to produce fBm and fGn series x(t) based on the characteristics of the power spectral density S(f) of these signals. In other words, the idea of SSM is to generate a right kind of S(f) that gives rise to fBm or fGn with an exponent 0 <H< 1.

Since S(f) is obtained as the square of the modulus of the Fourier transform of the signal and its proportional to fβ (see equation (8)),

Sf=Xf2fβ. (21)

We need to generate a complex series X(f) with a modulus r(f) proportional to 1/fβ/2. r(f) is obtained as follows:

rf=Wgnffβ/2, (22)

where Wgn is a white Gaussian noise and f is the frequency ranging from 1 to N/2.

To generate the X(f), we also need to generate a random phase in radian:

ϕf=2πWun, (23)

where Wun is a white noise with uniform distribution.

We can create the complex coefficients:

af=rfcosϕf,bf=rfsinϕf. (24)

These coefficients are extended to the whole range of expected values (in respect of the Shannon theorem), a(f) from (N/2)+ 1 to N being equal to a(f) from (N/2) to 1 and b(f) from (N/2)+ 1 to N being equal to b(f) from (N/2)to 1. Then, the complex series X(f) are generated as follows:

Xf=af+ibf. (25)

Finally, the inverse Fourier transform of this complex number is computed to obtain the time series:

xt=1Xf=1Nf=1NXfei2πft/N, (26)

where t = 1, 2,…, N.

3.3. ARFIMA Simulation Method

We used the Matlab code proposed by Fatichi [14], based on equation (12), and in the present case, as we limited ourselves to ARFIMA (0, d, 0) models, on equation (13). This code just bounds the summation in equation (15) to k = 100. We used this procedure for generating fGn series (d ∈ ]−0.5, 0.5[). In order to obtain the whole set of series we needed, we computed fBm series by cumulated summation.

We now present the three estimation methods we used. Note that for a better readability, the exponents of the simulated series and the estimates are expressed or converted in α metrics. Estimates are denoted with a circumflex (α^, β^, d^).

3.4. Detrended Fluctuation Analysis

The DFA algorithm works as follows, for a series x(t) of length N where t = 1, 2,…, N. The series is first integrated, by computing for each t the accumulated departure from the mean of the whole series:

Xt=i=1txix¯. (27)

This integrated series is then divided into k nonoverlapping intervals of length n. The last Nkn data points are excluded from analysis. Within each interval, a least squares line is fitted to the data. The series X(t) is then locally detrended by subtracting the theoretical values Xth(t) given by the regression. For a given interval length n, the characteristic size of fluctuation for this integrated and detrended series is calculated by the following equation:

Fn=1Nknt=1NknXtXTht2. (28)

In the original algorithm, this computation is repeated over all possible interval lengths, for example, from n = 10 to n = N/2 [9]. In the present paper, we applied the evenly spaced averaged version of DFA [29], which significantly reduces the variability of estimates. This procedure consists in dividing the (log) abscissa into P bins of length (log10(nmax/nmin))/P, starting from log10(nmin). The P bins are defined as follows:

binp=log10nmin+p1Plog10nmaxnmin;log10nmin+pPlog10nmaxnmin, (29)

where p = 1, 2,…, P. In the present analyses, we set nmin = 10, nmax = 512, and P = 18.

Within each bin p, the average interval length n¯p and the average fluctuation size F¯np are computed. A power law is expected as

F¯npn¯pα. (30)

The exponent estimate (α^) is obtained as the slope of the double logarithmic plot of F¯np as a function of n¯p.

3.5. Power Spectral Density Analysis (PSD)

This method works on the basis of the periodogram obtained by the fast Fourier transform algorithm and exploits the power law given by equation (8). β^ is estimated by calculating the negative slope (−β) of the line relating log (S(f)) to log f.

In the present paper, we used the improved version proposed by Fougere [30] and modified by Eke et al. [10], designated as lowPSDwe. This method uses a combination of preprocessing operations: first, the mean of the series is subtracted from each value, and then, a parabolic window is applied—each value in the series is multiplied by the following function:

Wt=12tN+112,xwt=xt×Wt, (31)

where t = 1, 2,…, N.

Secondly, a bridge detrending is performed by subtracting from the data the line connecting the first and last point of the series:

lt=xw1xwN×t1N1+xwN,xw,lt=xwtlt, (32)

where t = 1, 2,…, N.

The Fourier transform is applied on the modified series xw,l, and the fitting of β excludes the high-frequency power estimates (f > 1/8 of maximal frequency). This method was proven by Eke et al. [10] to provide more reliable estimates of the spectral index. For allowing a direct comparison between methods, β was then converted into α metrics by a simple linear transform α^=β^+1/2.

3.6. ARFIMA Modeling

The d parameter was estimated by fitting an ARFIMA(0, d, 0) model to the series using Whittle approximation of the maximum likelihood estimator. We used the ARFIMA(p, d, q) estimator package for Matlab proposed by Inzelt [31] on the Matlab central file exchange platform.

As previously explained, ARFIMA modeling holds only for fGn series and d is bounded within the interval ]−0.5, 0.5[. For fBm series, Diebolt and Guiraud [32] proposed to apply ARFIMA modeling to the corresponding fGn (obtained by differentiation) and then to estimate the theoretical fractional parameter of the fBm series by adding 1 to the d value obtained from the fGn. This strategy, however, requires an a priori assessment of the fGn/fBm classification of series. Some solutions, based on the preliminary application of methods working indifferently on fGn and fBm, such as DFA or PSD, have been proposed but yielded important percentages of misclassification around the 1/f boundary [9, 10].

In order to apply ARFIMA modeling indifferently on fGn and fBm series, as DFA and PSD, we used the following strategy: the algorithm consists in finding the best Whittle approximate of the maximum likelihood estimator by constrained optimization. According to ARFIMA properties, the output parameter is bounded, yielding the parameter value d^=0.4999 if the algorithm did not converge on the upper bound. When d^o, obtained from the original series, was equal to 0.4999, the series was considered as fBm and the algorithm was applied on the differentiated time series in order to obtain d^diff estimate. In this nonstationary case, d^=d^diff+1 else d^=d^o. This parameter estimate d^ was then converted into α metrics by a simple linear transform α^=2d^+1/2.

4. Results

We present in Figure 1 the relationships between the true α and the mean α^ values, for each simulation and estimation method. If we consider the global shape of results around the identity line, it seems obvious that estimation methods performed better when series were generated by the method belonging to the same domain (see the three graphs ranging along the descending diagonal). This was already observed by Eke et al. [11].

Figure 1.

Figure 1

Relationship between true and mean estimated α values. Top row: series generated by the DH algorithm; middle row: series generated by SSM; bottom row: series generated by ARFIMA simulation. Left column: estimations performed with DFA; center column: estimations performed with PSD; right column: estimations performed with ARFIMA modeling. Dashed lines represent identity.

The first row depicts the results obtained by the three estimation methods with the series generated by the DH algorithm. A common bias appears in the three graphs, with a strong underestimation of α in fBm series, close to the 1/f boundary. This reveals a clear disruption in the original fGn/fBm model. This phenomenon was not observed when series were generated with the two other methods.

DFA tends to underestimate α for fBm series, especially when series were generated by SSM or ARFIMA modeling. Conversely, PSD tends to overestimate α in fBm series, especially when series were generated by the DH algorithm or by ARFIMA modeling.

Figure 2 represents a focus of the previous results on the 1/f boundary (i.e., from α = 0.9 to α = 1.1). The first row confirms the clear disruption, around the 1/f boundary, for series generated with the DH algorithm. Additionally, it reveals a slight overestimation in fGn series, in the vicinity of 1/f noise, for PSD and ARFIMA modeling. In contrast, series generated by SSM and ARFIMA simulation present a clear continuity around the 1/f boundary, whatever the method of estimation. DFA tends to overestimate α for series generated by spectral synthesis, but not for series simulated with ARFIMA. PSD tends to overestimate α in this particular range. Finally, ARFIMA modeling slightly overestimates α for fBm series generated by SSM.

Figure 2.

Figure 2

Relationship between true and mean estimated α values, with a focus on the 1/f boundary (0.9 ≤ α ≥ 1.1). Top row: series generated by the DH algorithm; middle row: series generated by SSM; bottom row: series generated by ARFIMA simulation. Left column: estimations performed with DFA; center column: estimations performed with PSD; right column: estimations performed with ARFIMA modeling. Dashed lines represent identity.

Finally, we present in Figure 3 the variability (standard deviation) of the α^ samples, for each simulation and estimation method. These graphs are presented with the same scale on the vertical axe, in order to facilitate comparisons between methods. Globally, α^ appears less variable in series generated by the DH algorithm or with ARFIMA simulation, than in those obtained from SSM, which yielded the worst results. Concerning estimation methods, DFA results show a clear increase in α^ variability as true α increases. This effect is not present with PSD and ARFIMA modeling, in which variability remains stable over the whole true α range. PSD results reveal a high level of variability, and especially when series were generated by SSM. Conversely, ARFIMA modeling is characterized by a low variability.

Figure 3.

Figure 3

Relationship between true α and the variability of estimated α values (standard deviation). Top row: series generated by the DH algorithm; middle row: series generated by SSM; bottom row: series generated by ARFIMA simulation. Left column: estimations performed with DFA; center column: estimations performed with PSD; right column: estimations performed with ARFIMA modeling.

5. Discussion

5.1. Davies–Harte Algorithm

The most important observation with the series generated with the DH algorithm is the strong bias for fBm series close to the 1/f boundary. This result, consistently obtained with the three analysis methods, should obviously be attributed to a bias in the generation technique. This result was already evidenced by Stadnitski [27], who suggested that “the observed discrepancies probably occurred due to the incapability of the Davies–Harte technique to provide fBm series with H close to 0. These results underline the importance of a proper data generation in simulation studies and indicate a revision of results from studies that employed the Davies–Harte technique” (pp. 144‐145).

Delignières [28] showed that this bias was not related to the Davies–Harte algorithm, but more fundamentally to the fGn/fBm model itself. Based on the premises of the model, and especially on the autocorrelation of fGn (equation (5)), the author derived an analytical expression for the autocorrelation of fBm:

ρfBmτ=1NN1τ2H2t=1N1Ntt2H, (33)

where N is the length of the series. We report in Figure 4 the theoretical lag-1 autocorrelations, computed for fGn series ranging from H = 0 to 1 according to equation (5) and for fBm series ranging from H = 0 to 1 according to equation (33), considering N = 1024. These results evidence a severe breakdown of the correlation properties of series around the fGn/fBm boundary, and the limit behavior of fBm, H approaching 0, is uncorrelated white noise.

Figure 4.

Figure 4

Theoretical lag-1 autocorrelation for fGn (a), based on equation (5), and for fBm (b), based on equation (33), for α values ranging from 0 to 2 (adapted from Delignières [28]).

This suggests that fBm series, obtained by the cumulative summation of the corresponding fGn series, should in fact be fGn for very low H values. On the basis of equation (33), we computed the expected ρfBm(1) values, for H ranging from 0.01 to 0.1, and we compared these expected values to the mean lag-1 autocorrelations observed in the series simulated with the Davies–Harte algorithm. The results are illustrated in Figure 5 and show that the correlational structure of simulated series matched closely that expected from the fGn/fBm model.

Figure 5.

Figure 5

Expected and observed lag-1 autocorrelation for fBm signals. Solid line: theoretical lag-1 autocorrelation of fBm signals, for H ranging from 0.01 to 0.1, computed on the basis of equation (33). White circles: mean lag-1 autocorrelations observed in the series simulated with the Davies–Harte algorithm.

5.2. Spectral Synthesis Method

In contrast with the DH algorithm, SSM provides continuity around the 1/f boundary. This continuity was expected, as SSM works indifferently over the whole range of β exponents. This result could be surprising, as fGn/fBm and spectral models are often considered equivalent, representing similar properties in the time and frequency domains, respectively.

Another important result is related to the variability of estimates, which appears systematically higher in series generated by SSM than that observed with other methods of generation. This represents an important problem with this method, which seems unable to provide series sufficiently close to the expected exponents.

5.3. ARFIMA Simulation

ARFIMA simulation also provides a nice continuity around the 1/f boundary. This result is interesting, as fBm series were in that case obtained by cumulative summation of their correspondent fGn, as for the DH algorithm. Additionally, the obtained sets of series presented the lowest variability in exponent estimations, whatever the used method. These results suggest that one could get a good confidence in series generated by ARFIMA, as compared with the two other methods.

5.4. Detrended Fluctuation Analysis

DFA works quite well with fGn series but presents a systematic negative bias and a high level of variability for fBm series. For understanding these poor results with fBm series, as compared with other methods, it is important to keep in mind that DFA actually works on integrated series and in this case on integrated fBm. This family of overdiffusive processes is not well known, and the diffusion property exploited by DFA seems moderately appropriate with such series [9].

DFA is a very popular method, which presents the advantage to be indifferently applicable to both fGn and fBm. Some other methods gave satisfying results but are only relevant for fGn (e.g., the dispersional analysis or the rescaled range analysis [33]) or for fBm (e.g., the scaled windowed variance analysis [34]). However, researchers are often unable to know if their signals refer to any of these families, especially if these signals are close to the 1/f boundary. DFA allows overcoming this problem. The poor performances of DFA with fBm signals could be neglected, considering that most physiological series fall into the fGn family. In some cases, however, for example, for postural sway or gaze fluctuations, series are clearly nonstationary and should be considered as fBm. In such cases, one could consider to apply DFA on differenced series or to omit the integration step in the algorithm (see, e.g., [35, 36]).

It is worth noting that the present results were obtained with the evenly spaced version of DFA, which was proven to significantly improve the accuracy and to reduce the variability of the original method [29]. Results would have been even worse with the original algorithm.

5.5. Power Spectral Density

Despite the application of the refinements proposed by Eke et al. [10], PSD provided the worst estimation results. In terms of accuracy, PSD tends to overestimate α in fBm series, and results are particularly deceptive in terms of variability. PSD remains a quite popular method, especially because it provides appealing graphical results. The bilogarithmic representation of the power spectrum, beyond the estimation of the 1/f slope, can give essential indications about the presence and the nature of short-term fluctuations in the series (see, e.g., [37, 38]). However, for the specific purpose of exponent estimation, PSD seemed clearly outperformed by other methods.

5.6. ARFIMA Modeling

ARFIMA modeling has been neglected by most previous studies comparing fractal methods [911]. Rangarajan and Ding [39] developed an integrated approach of fractal analysis associating rescaled range and spectral analyses, but they never considered ARFIMA as a possible alternative or complement. However, this method provided the most accurate and less variable estimates. As indicated in the introduction, ARFIMA modeling was initially designed for accounting for stationary series. As proposed by Diebolt and Guiraud [32], we differentiated nonstationary series before the application of ARFIMA modeling, and this method gave satisfying results. We applied a very simple procedure for distinguishing stationary and nonstationary series, considering that series yielding d estimates equals to the upper limit obtainable with the algorithm (0.49999) should be considered nonstationary. This simple procedure provided good results, as evidenced by the nice continuity observed around the 1/f boundary. In the initial steps of this work, we tested the ARFIMA package [40, 41] for the matrix computing language Ox [42]. This algorithm, however, gave worse results around the 1/f boundary. Liu et al. [43] recently proposed a first evaluation of diverse ARFIMA programs, available on various platforms (Matlab, R, SAS, and OxMetrics), providing solutions for simulation, parameter estimation, and forecasting. Further investigations should be necessary for providing effective guidelines for the selection of the most relevant solutions.

In the present paper, we limited ourselves to the estimation of ARFIMA (0, d, 0) models. ARFIMA modeling could also account for various autoregressive and/or moving average processes that could contaminate empirical series. This method allows isolating long-range correlations in series and then to provide a better estimation of fractal exponents. Finally, a procedure using ARFIMA modeling has been proposed for gauging the effective presence of genuine long-range correlations in empirical series [26, 38, 44]. Indeed, short-term processes could sometimes mimic 1/f-like fluctuations, yielding false detections of long-range correlations with classical methods such as DFA [38]. The method proposed by Torre et al. [26] allows testing the relative likelihood of various ARMA and ARFIMA models and finally to conclude to the effective presence of long-range correlations.

In order to illustrate the experimental usefulness of these results, we applied the three estimation methods to a set of series of stride durations, collected during 15 min walking bouts with two groups of participants: the first one was composed of 22 young participants (28.07 yrs ± 8.88) and the second of 23 older persons (72.36 yrs ± 4.88). These data were collected during recent experiments performed in our lab [45, 46], and each series had a length of 512 data points. As indicated previously, Hausdorff et al. [47] showed that aging was characterized by a typical loss of complexity in walking dynamics, and one could expect to evidence significant differences in α estimates between the two groups. We present in Table 1 the results obtained with the three methods, converted in α metrics for comparison. In accordance with our previous results, PSD yielded a greater variability in estimates than DFA or ARFIMA. We compared the two samples with a one-way ANOVA, which did not evidence any difference between the two groups on the basis of PSD estimates. In contrast, a significant difference was found for DFA and ARFIMA estimates, with a stronger effect size for ARFIMA.

Table 1.

Mean α estimates (standard deviation in parentheses), computed with the three methods (DFA, PSD, and ARFIMA), from series of series of stride durations, collected during 15 min walking bouts with young (N = 22) and older (N = 23) participants.

DFA PSD ARFIMA
Young Mean (αDFA) SD 0.878 (0.096) 1.023 (0.137) 0.887 (0.096)
Elderly Mean (αDFA) SD 0.783 (0.112) 0.951 (0.196) 0.764 (0.088)
F (1, 43) 9.397 1.999 20.044
p 0.004 0.165 0.000
Partial η2 0.179 0.044 0.318

Beyond our observations about the superiority of ARFIMA modeling, in terms of accuracy and variability of exponent estimation on exact synthetic signals, this final result shows that ARFIMA modeling provides a better statistical power than DFA or PSD in the analysis of experimental data [48]. Further investigations remain necessary, however, for assessing the respective performances of these methods, especially with shorter time series, which represent an essential challenge in human behavioral experiments [912, 21, 29, 49].

6. Conclusion

This study provides a strong support for the use of ARFIMA methods, for simulation as well as for parameters estimation purposes. As previously indicated, ARFIMA methods were not considered in recent comparative studies and rarely exploited in empirical works, at least in the physiological and behavioral domains. The accuracy and the low variability of exponent estimation with ARFIMA modeling should convince researchers of the advantages of this method, especially for detecting mean differences between groups. This method should attract the attention of researchers, either for their future experiments or for revisiting their previous results.

Acknowledgments

This work was supported by the University of Montpellier-France (Grant no. BUSR-2014). The authors thank Dr Zainy M.H. Almurad for providing them with the data used in Table 5.

Data Availability

The time series used to support the findings of this study are available from the corresponding author upon request.

Disclosure

This work has been partly presented as a poster during the 13th Annual Meeting of the Doctoral School 463 “Human Movement Sciences,” Montpellier, June 2, 2017, and during the 17th International Congress of the Association for Research in Physical Activities and Sports (ACAPS), Dijon, October 29–31, 2017.

Conflicts of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as potential conflicts of interest.

References

  • 1.Gilden D., Thornton T., Mallon M. 1/f noise in human cognition. Science. 1995;267(5205):1837–1839. doi: 10.1126/science.7892611. [DOI] [PubMed] [Google Scholar]
  • 2.Hausdorff J. M., Peng C. K., Ladin Z., Wei J. Y., Goldberger A. L. Is walking a random walk? Evidence for long-range correlations in stride interval of human gait. Journal of Applied Physiology. 1995;78(1):349–358. doi: 10.1152/jappl.1995.78.1.349. [DOI] [PubMed] [Google Scholar]
  • 3.Lemoine L., Torre K., Delignières D. Testing for the presence of 1/f noise in continuation tapping data. Canadian Journal of Experimental Psychology/Revue Canadienne de Psychologie Expérimentale. 2006;60(4):247–257. doi: 10.1037/cjep2006023. [DOI] [PubMed] [Google Scholar]
  • 4.Peng C.-K., Mietus J., Hausdorff J. M., Havlin S., Stanley H. E., Goldberger A. L. Long-range anticorrelations and non-Gaussian behavior of the heartbeat. Physical Review Letters. 1993;70(9):1343–1346. doi: 10.1103/physrevlett.70.1343. [DOI] [PubMed] [Google Scholar]
  • 5.Hausdorff J. M. Gait dynamics in Parkinson’s disease: common and distinct behavior among stride length, gait variability, and fractal-like scaling. Chaos: An Interdisciplinary Journal of Nonlinear Science. 2009;19(2) doi: 10.1063/1.3147408.026113 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Hunt N., McGrath D., Stergiou N. The influence of auditory-motor coupling on fractal dynamics in human gait. Scientific Reports. 2014;4(1):p. 5879. doi: 10.1038/srep05879. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Kaipust J. P., McGrath D., Mukherjee M., Stergiou N. Gait variability is altered in older adults when listening to auditory stimuli with differing temporal structures. Annals of Biomedical Engineering. 2013;41(8):1595–1603. doi: 10.1007/s10439-012-0654-9. [DOI] [PubMed] [Google Scholar]
  • 8.Rhea C. K., Kiefer A. W., D’Andrea S. E., Warren W. H., Aaron R. K. Entrainment to a real time fractal visual stimulus modulates fractal gait dynamics. Human Movement Science. 2014;36:20–34. doi: 10.1016/j.humov.2014.04.006. [DOI] [PubMed] [Google Scholar]
  • 9.Delignieres D., Ramdani S., Lemoine L., Torre K., Fortes M., Ninot G. Fractal analyses for ‘short’ time series: a re-assessment of classical methods. Journal of Mathematical Psychology. 2006;50(6):525–544. doi: 10.1016/j.jmp.2006.07.004. [DOI] [Google Scholar]
  • 10.Eke A., Herman P., Bassingthwaighte J. B., et al. Physiological time series: distinguishing fractal noises from motions. Pflügers Archiv: European Journal of Physiology. 2000;439(4):403–415. doi: 10.1007/s004240050957. [DOI] [PubMed] [Google Scholar]
  • 11.Eke A., Herman P., Kocsis L., Kozak L. R. Fractal characterization of complexity in temporal physiological signals. Physiological Measurement. 2002;23(1):R1–R38. doi: 10.1088/0967-3334/23/1/201. [DOI] [PubMed] [Google Scholar]
  • 12.Jennane R., Harba R., Jacquet G. Analysis methods for fractional Brownian motion: theory and comparative results. Trait Signal. 2001;18:419–436. [Google Scholar]
  • 13.Davies R. B., Harte D. S. Tests for Hurst effect. Biometrika. 1987;74(1):95–101. doi: 10.1093/biomet/74.1.95. [DOI] [Google Scholar]
  • 14.Fatichi S. ARFIMA simulations. 2009. https://www.mathworks.com/matlabcentral/fileexchange/25611-arfima-simulations.
  • 15.Saupe D. Algorithms for random fractals. In: Peitgen H. O., Saupe D., editors. The Science of Fractal Images. Berlin, Germany: Springer; 1988. pp. 71–136. [Google Scholar]
  • 16.Mandelbrot B. B., Van Ness J. W. Fractional Brownian motions, fractional noises and applications. SIAM Review. 1968;10(4):422–437. doi: 10.1137/1010093. [DOI] [Google Scholar]
  • 17.Voss R. F. 1/f (flicker) noise: a brief review. Proceedings of the 33rd Annual Symposium on Frequency Control; May-June 1979; Atlantic City, NJ, USA. pp. 40–46. [DOI] [Google Scholar]
  • 18.Granger C. W. J., Joyeux R. An introduction to long-memory time series models and fractional differencing. Journal of Time Series Analysis. 1980;1(1):15–29. doi: 10.1111/j.1467-9892.1980.tb00297.x. [DOI] [Google Scholar]
  • 19.Taqqu M. S., Teverovsky V., Willinger W. Estimators for long-range dependence: an empirical study. Fractals. 1995;3(4):785–798. doi: 10.1142/S0218348X95000692. [DOI] [Google Scholar]
  • 20.Flandrin P. On the spectrum of fractional Brownian motions. IEEE Transactions on Information Theory. 1989;35(1):197–199. doi: 10.1109/18.42195. [DOI] [Google Scholar]
  • 21.Delignières D., Marmelat V. Theoretical and methodological issues in serial correlation analysis. Advances in Experimental Medicine and Biology. 2013;782:127–148. doi: 10.1007/978-1-4614-5465-6_7. [DOI] [PubMed] [Google Scholar]
  • 22.Box G. E. P., Jenkins G., Reinsel G. C., Ljung G. M. Time Series Analysis: Forecasting and Control. San Francisco, CA, USA: Holden Day; 1976. [Google Scholar]
  • 23.Fortes M., Ninot G., Delignières D. The auto-regressive integrated moving average procedures: implications for adapted physical activity research. Adapted Physical Activity Quarterly. 2005;22(3):221–236. doi: 10.1123/apaq.22.3.221. [DOI] [Google Scholar]
  • 24.Beran J. Statistics for Long-Memory Processes. Boca Raton, FL, USA: CRC Press; 1994. [Google Scholar]
  • 25.Beran J., Feng Y., Ghosh S., Kulik R. Long-Memory Processes. Berlin, Germany: Springer; 2013. [DOI] [Google Scholar]
  • 26.Torre K., Delignières D., Lemoine L. Detection of long-range dependence and estimation of fractal exponents through ARFIMA modelling. British Journal of Mathematical and Statistical Psychology. 2007;60(1):85–106. doi: 10.1348/000711005X89513. [DOI] [PubMed] [Google Scholar]
  • 27.Stadnitski T. Some critical aspects of fractality research. Nonlinear Dynamics, Psychology, Life Sciences. 2012;16:137–158. [PubMed] [Google Scholar]
  • 28.Delignières D. Correlation properties of (discrete) fractional Gaussian noise and fractional Brownian motion. Mathematical Problems in Engineering. 2015;2015:7. doi: 10.1155/2015/485623.485623 [DOI] [Google Scholar]
  • 29.Almurad Z. M. H., Delignières D. Evenly spacing in detrended fluctuation analysis. Physica A: Statistical Mechanics and Its Applications. 2016;451:63–69. doi: 10.1016/j.physa.2015.12.155. [DOI] [Google Scholar]
  • 30.Fougere P. F. On the accuracy of spectrum analysis of red noise processes using maximum entropy and periodogram methods: simulation studies and application to geophysical data. Journal of Geophysical Research. 1985;90(A5):4355–4366. doi: 10.1029/JA090iA05p04355. [DOI] [Google Scholar]
  • 31.Inzelt G. Maximum Likelihood Estimators of Stationary Univariate ARFIMA (p, d, q) Processes. 2011. https://fr.mathworks.com/matlabcentral/fileexchange/30238-arfima-p-d-q-estimator. [Google Scholar]
  • 32.Diebolt C., Guiraud V. A note on long memory time series. Quality and Quantity. 2005;39(6):827–836. doi: 10.1007/s11135-004-0436-z. [DOI] [Google Scholar]
  • 33.Caccia D. C., Percival D., Cannon M. J., Raymond G., Bassingthwaighte J. B. Analyzing exact fractal time series: evaluating dispersional analysis and rescaled range methods. Physica A: Statistical Mechanics and Its Applications. 1997;246(3-4):609–632. doi: 10.1016/S0378-4371(97)00363-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.Cannon M. J., Percival D. B., Caccia D. C., Raymond G. M., Bassingthwaighte J. B. Evaluating scaled windowed variance methods for estimating the Hurst coefficient of time series. Physica A: Statistical Mechanics and Its Applications. 1997;241(3-4):606–626. doi: 10.1016/S0378-4371(97)00252-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35.Delignières D., Torre K., Bernard P.-L. Transition from persistent to anti-persistent correlations in postural sway indicates velocity-based control. PLoS Computational Biology. 2011;7(2) doi: 10.1371/journal.pcbi.1001089.e1001089 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 36.Stephen D. G., Anastas J. Fractal fluctuations in gaze speed visual search. Attention, Perception, and Psychophysics. 2011;73(3):666–677. doi: 10.3758/s13414-010-0069-3. [DOI] [PubMed] [Google Scholar]
  • 37.Delignières D., Lemoine L., Torre K. Time intervals production in tapping and oscillatory motion. Human Movement Science. 2004;23(2):87–103. doi: 10.1016/j.humov.2004.07.001. [DOI] [PubMed] [Google Scholar]
  • 38.Wagenmakers E.-J., Farrell S., Ratcliff R. Estimation and interpretation of 1/fα noise in human cognition. Psychonomic Bulletin and Review. 2004;11(4):579–615. doi: 10.3758/BF03196615. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 39.Rangarajan G., Ding M. Integrated approach to the assessment of long range correlation in time series data. Physical Review E. 2000;61(5):4991–5001. doi: 10.1103/PhysRevE.61.4991. [DOI] [PubMed] [Google Scholar]
  • 40.Doornik J. A., Ooms M. A Package for Estimating, Forecasting, and Simulating Arfima Models: Arfima Package 1.0 for Ox. Rotterdam, Netherlands: Erasmus University; 1999. [Google Scholar]
  • 41.Ooms M., Doornik J. A. Estimation, Simulation and Forecasting for Fractional Autoregressive Integrated Moving Average Models. Proceedings of the Fourth International Symposium in Computational Economics and Finance; 1998; Cambridge, UK. [Google Scholar]
  • 42.Doornik J. A. Ox: An Object-Oriented Matrix Language. London, UK: Timberlake Consultants Press; 2001. [Google Scholar]
  • 43.Liu K., Chen Y., Zhang X. An evaluation of ARFIMA (autoregressive fractional integral moving average) programs. Axioms. 2017;6(4):p. 16. doi: 10.3390/axioms6020016. [DOI] [Google Scholar]
  • 44.Wagenmakers E.-J., Farrell S., Ratcliff R. Human cognition and a pile of sand: a discussion on serial correlations and self-organized criticality. Journal of Experimental Psychology: General. 2005;134(1):108–116. doi: 10.1037/0096-3445.134.1.108. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 45.Almurad Z. M. H., Roume C., Blain H., Delignieres D. Complexity matching: restoring the complexity of locomotion in older people through arm-in-arm walking. Frontiers in Physiology. 2018;9:p. 1766. doi: 10.3389/fphys.2018.01766. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 46.Almurad Z. M. H., Roume C., Delignières D. Complexity matching in side-by-side walking. Human Movement Science. 2017;54:125–136. doi: 10.1016/j.humov.2017.04.008. [DOI] [PubMed] [Google Scholar]
  • 47.Hausdorff J. M., Mitchell S. L., Firtion R., et al. Altered fractal dynamics of gait: reduced stride-interval correlations with aging and Huntington’s disease. Journal of Applied Physiology. 1997;82(1):262–269. doi: 10.1152/jappl.1997.82.1.262. [DOI] [PubMed] [Google Scholar]
  • 48.Kuznetsov N. A., Rhea C. K. Power considerations for the application of detrended fluctuation analysis in gait variability studies. PLoS One. 2017;12(3) doi: 10.1371/journal.pone.0174144.e0174144 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 49.Delignieres D., Marmelat V. Fractal fluctuations and complexity: current debates and future challenges. Critical Reviews in Biomedical Engineering. 2012;40(6):485–500. doi: 10.1615/critrevbiomedeng.2013006727. [DOI] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

The time series used to support the findings of this study are available from the corresponding author upon request.


Articles from Computational and Mathematical Methods in Medicine are provided here courtesy of Wiley

RESOURCES