Skip to main content
PLOS One logoLink to PLOS One
. 2022 Jul 22;17(7):e0271969. doi: 10.1371/journal.pone.0271969

Testing for Benford’s Law in very small samples: Simulation study and a new test proposal

Andrea Cerasa 1,*
Editor: Miguel A Fernández2
PMCID: PMC9307211  PMID: 35867697

Abstract

Benford’s Law defines a statistical distribution for the first and higher order digits in many datasets. Under very general condition, numbers are expected to naturally conform to the theorized digits pattern. On the other side, any deviation from the Benford distribution could identify an exogenous modification of the expected pattern, due to data manipulation or even fraud. Many statistical tests are available for assessing the Benford conformity of a sample. However, in some practical applications, the limited number of data to analyze may raise questions concerning their reliability. The first aim of this article is then to analyze and compare the behavior of Benford conformity testing procedures applied to very small samples through an extensive Monte Carlo experiment. Simulations will consider a thorough choice of compliance tests and a very heterogeneous selection of alternative distributions. Secondly, we will use the simulation results for defining a new testing procedure, based on the combination of three tests, that guarantees suitable levels of power in each alternative scenario. Finally, a practical application is provided, demonstrating how a sounding testing Benford compliance test for very small samples is important and profitable in anti-fraud investigations.

Introduction

Benford’s Law (BL) defines a probability distribution for patterns of significant digits in numerical data. Its formulation is grounded on the intriguing observation made first by Newcomb [1], then by Benford [2], who noticed a non-uniform amounts of wear in the pages of the logarithmic tables. In its complete form, the law states that the leading digits of many natural phenomena are not uniformly distributed, as one may expect, but follow a logarithmic distribution:

P[D1(X)=d1,,Dm(X)=dm]=log10(1+1l=1m10m-ldl) (1)

where Dj(x) is the j-th significant digit of a positive real number x, d1 ∈ {1, …, 9}, and dj ∈ {0, …, 9} for j = 2, …, m. Focusing only on the first significant digit (FSD), expression (1) reduces to:

P[D1(X)=k]=pkB=log10(1+1/k),fork=1,2,,9. (2)

Theoretical studies investigated the properties and provided a limit theorem for the digit distribution. Classical references for this topic are [36]. At the same time, empirical applications have shown that many sets of numerical data are consistent with BL, at least in its simplest form Eq (2). Some examples in this sense are stock indexes [7], hydrology data [8] and volcanology data [9]. In addition, thanks to its generality and feasibility in different fields, the Benford distribution has fruitfully supported fraud investigations and the detection of manipulated data, in particular concerning COVID declared figures [10], scientific studies [11], media and social networks data [12] and international trade [13]. The assumption is that clean data (i.e. without any external manipulation) are distributed according to Eq (1). Generally, this condition is satisfied whenever numbers are the result of mathematical operations (multiplication, division, raising to power and so on) on values taken from different random variables, as in the case of accounting data [14]. The value of a purchase, for example, is the outcome of the multiplication between the number of items and their unitary price, which itself comes from the combination of different numbers. In such context, a significant departure from the theoretical distribution can point to data sets that include fabricated numbers.

The identification of non-Benford numbers can rely on a variety of statistical tests. Their empirical properties may significantly differ, but we expect that their power, ceteris paribus, increases with the sample dimension. On the contrary, when the number of observations is very small, “there may be insufficient power to meaningfully detect or confirm conformance with the law” ([15], page 2793). Nevertheless, in many practical applications, the number of figures available for each individual sample to test could be quite limited. In this case, the usual solution for increasing the expected sensitivity is to run the BL compliance tests on several individual samples merged together. Consider, for example, the investigations aimed to assess the digit distribution of the numbers published in scientific journals. Their analyses are usually not performed article by article, but they rather consider groups of articles pulled together (for example, those published in the same year, as in [16]). This strategy is not feasible however when the aim is to identify the specific individual sample that may contain irregularities. This is just the case of anti-fraud, where the target is to identify the economic operators that may have manipulated their declared numbers. Thus, in such contexts, the only strategy is to maximize the reliability of the expected outcome, that is to choose the testing procedure that, for a given significance level, guarantees the largest expected power against a wide range of possible alternatives.

In this article, we firstly analyze and compare the behavior of several testing procedures against a huge set of alternative scenarios. Besides considering a thorough choice of Benford compliance tests and a very heterogeneous selection of alternative distributions, the main novelty with respect to previous works with similar aim (as, for example [1719]), is that we focus on samples with very small dimensions (i.e. n = 20). We do not expect to find a testing procedure that strictly dominates the other in all the different scenarios. The target is rather to find the ones that show the best general behavior independently of the alternative. Secondly, we use the testing performances obtained in the simulation exercise to derive and propose a combined test that guarantees suitable levels of power in each alternative scenario. The objective of this study is then twofold. From one side, we provide a comprehensive analysis of small samples properties of Benford’s compliance tests that could support the researchers in the selection of the proper procedure. From the other, we propose an alternative method based on the combination of different tests that offers desirable performance against a wide range of possible deviations. An empirical application on international trade data shows how the availability of reliable BL testing procedures for very small samples can remarkably increase the range of applicability, and provide a valuable support for limiting the number of economic transactions that deserve further anti-fraud investigations.

The paper proceeds as follows. In the next section, the set of BL tests compared in the simulation exercise are briefly introduced, properly divided into three families: tests for the FSD, tests for the complete form of the BL and the summation test. This is followed by a section that describes the different alternative distributions of values considered in the Monte Carlo experiment. Therefore, the simulation results are presented and discussed. In addition, we used them to define a combined test with desirable empirical properties. We present an empirical application of Benford compliance tests on international trade data, stressing the importance of reliable methods for small samples in such context. The final section concludes.

Testing the BL conformity

Many testing procedures allow to assess the conformity of a set of values to the Benford’s theory. In the first part of our study, we want to study their behavior in small samples through a huge simulation exercise. We took into account most of the testing procedures proposed in past investigations, in order to provide a complete comparison. They can be divided into three families: tests for the FSD, tests for the complete BL, and the summation tests recently proposed by [20].

Tests for the FSD

Given an n-sample of values {x1, …, xn}, we can assess the BL conformity of the FSD through the null hypothesis:

H0:pk=log10(1+1/k),fork=1,2,,9 (3)

where pk = nk/n, and nk=i=1nI(D1(xi)=k).

The first test is the well known Pearson goodness-of-fit test:

χ2=k=19(nk-npkB)2npkB=nk=19(pk-pkB)2pkB. (4)

that is asymptotically distributed as a χ82. It is surely one of most used statistics in BL compliance experiments, even though its potential low power, especially in small samples, suggests some cautions [21]. In this simulation exercise, it could be considered as a benchmark for assessing the gain in power associated with the other testing procedures.

A second category of FSD tests derives from the Cramer-Von-Mises, Watson and Anderson-Darling statistics, which are mainly used for testing the goodness-of-fit for continuous distributions [22, 23]. Let Fk=j=1kpk and FkB=j=1kpkB denote the cumulative distributions of respectively the empirical and the expected proportions, whose difference is defined as Zk=Fk-FkB. Define also the weights wk=(pkB+pk+1B)/2 for k = 1, …, 8 and w9=(p9B+p1B)/2, and use them for calculating the weighted mean of the cumulative distributions distances: Z¯=k=19wkZk. The three statistics are defined as:

W2=nk=19Zk2wkU2=nk=19(Zk-Z¯)2wkA2=nk=18Zk2wkFkB(1-FkB) (5)

Asymptotic critical values are available for the three tests [19].

Then, we will consider two tests based on the Kolmogorov-Smirnov (KS) distance of the cumulative FSD distributions. In particular, the first is simply the KS deviation:

KSd=nsup1k9|FkB-Fk|. (6)

whereas the second is the Kuiper test [24], defined as:

KUd=sup1k9(FkB-Fk)+sup1k9(Fk-FkB). (7)

Finally, the last FSD test considered is the Mean Test introduced by [16] and based on the simple observation that, if the FSD is distributed according to Eq (2), then its expected value is equal to 3.440 and its variance to 6.057. The Mean Test is defined as:

MT=|k=19pkdk-3.440|6.057. (8)

Tests for the complete form of the BL

The second family of tests applied in the simulation experiment does not limit the attention to the FSD, but considers the whole digit distribution (1). Defining the significand of a value yR as S(y)=10log10|y|log10|y|, with y=max{mZ:my} representing the floor function [5, 25], proved that:

YBenfordS(Y)10U

where U is a Uniform random variable on [0; 1[. Based on this theoretical result and ordering the observations {x1, …, xn} according to the value of the significand {S(x(1)) ≤ ⋯ ≤ S(x(n))}, the BL conformity could be tested through the statistics based on the KS distance applied on the logarithm of the ordered significands. As before, we considered both the KS test:

KSs=maxi[1,n]|log10S(x(i))-i/n| (9)

and the Kuiper test:

KUs=maxi[1,n][i/n-log10S(x(i))]+supi[1,n][log10S(x(i))-(i-1)/n]. (10)

Recently [26], showed that KSs test for uniformity provides very good results for small, medium size and even large records.

In addition, we also consider the Anderson-Darling test [27], defined as:

AD=i=1n2i-1n[ln(log10S(x(i)))+ln(1-log10S(x(n+1-i)))]. (11)

Summation test

Finally, the last test considered is the summation test Q [20]. Starting from the definition of the significand previously introduced, we define Zk(xi)=S(xi)×I(FSD(xi)=k) and Z¯k=n-11nZk(xi). According to these definitions and representing C = log10 e, the summation test limited to the first-digit case is given by:

Q=n(Z¯-μ)Σ-1(Z¯-μ) (12)

where Z¯=(Z¯1,,Z¯9), μ is the nine elements vector μ = (C, …, C)′, and Σ is the (9 × 9) matrix with elements σkk = C(k + 1/2 − C) and σkj = −C2 whenever kj.

Description of the alternative distributions

The finite sample behavior of the eleven testing procedures introduced in the previous section are investigated in a simulation study. Table 1 presents all the alternatives distributions considered in the simulation experiment, together with the corresponding parameter space. In each simulation s, a sample Xs={x1s,xns} is generated according to one of the pattern listed in the Table 1. Actually, the first three distributions are alternative patterns only for the FSD of x. In this case, the remaining digits of each number are simulated according to the Benford probabilities, in order to allow a suitable calculation of the tests for the complete form of the BL and of the summation test.

Table 1. Alternative distributions considered in the simulation experiment.

Family Parameter Space
GB θ ∈ [−1.5, 1.5]
R β ∈ [−14, 10]
H ρ ∈ (0, 10]
LN XLognormal(μ, σ) with μ ∈ {0, 0.5, 1, 1.5, 2}, σ ∈ [0.1, 1]
WB XWeibull(a, b) with a ∈ {0.5, 1, 1.5, 2, 2.5}, b ∈ [0.5, 5]

The first alternative family for the FSD is the Generalized Benford’s Law (GB, [28, 29]):

P[D1(X)=k]=pkGB(θ)={log10(1+1k)θ=0(k+1)θ-kθ10θ-1θ0 (13)

where k ∈ {1, 2, …, 9} and θR. Therefore, when θ = 0, the GB reduces to the standard Benford distribution for the FSD (2), whereas for θ = 1, the resulting distribution is uniform. This last case is quite interesting, since it mimics a manipulation strategy widely used in past applications [13, 19], based on a naive assignation of the first digit with equal probabilities. Finally, when θ < 0, more weight is assigned to smaller digits (1 and 2), whereas when θ > 1, more weight is assigned to bigger digits (from 4 to 9, see panel a of Fig 1).

Fig 1. FSD associated to the alternative distributions.

Fig 1

Then, we considered the FSD patterns proposed by Rodriguez (R, [30]):

P[D1(X)=k]=pkR(β)={19[1+109ln(10)+kln(k)-(k+1)ln(k+1)]β=0log10(1+1k)β=-1β+19β-(k+1)β+1-kβ+1β(10β+1-1)otherwise (14)

where k ∈ {1, 2, …, 9} and βR. Again, the standard Benford distribution (2) and the uniform distribution of the FSD are two particular cases of this family. The former corresponds to β = −1, whereas the latter to β = ±∞. In general, when β < 0 (β > 0), the digit probabilities are decreasing according to a convex (concave) pattern (see panel b of Fig 1).

The last FSD distributions considered are the ones proposed by Hürlimann (H, [31]):

P[D1(X)=k]=pkH(ρ)=12{[log10(1+k)]ρ-[log10k]ρ-[1-log10(1+k)]ρ+[1-log10k]ρ} (15)

with k ∈ {1, 2, …, 9} and ρ > 0. When ρ = 1 or ρ = 2, the distribution of the FSD corresponds to (2), whereas for other values of the parameter, the resulting distribution has a convex U-shape (see panel c of Fig 1).

The last two alternative distributions are the Lognormal (LN) and the Weibull (WB), two continuous random variables that may reasonably arise in many fields. Both of them can become very close to a Benford random variable for some combination of their parameters. Actually [4, 32], showed that a LN random variable with a large shape parameter is practically indistinguishable from a Benford random variable. This is why, in the simulation exercise, we decided to fix the scale parameter μ to five different values, and to let the shape parameter σ vary from 0.1 to 1. Similar arguments apply also to the WB random samples. In this case, the simulated values closely fit the Benford distribution for small values of the shape parameter [6]. Again, we fixed scale parameter a to five different values, and to let the shape parameter b vary from 0.5 to 5. Panels d and e of Fig 1 provide a representation of the FSD probabilities associated to the LN and WB distributions when the shape parameter is respectively 0.5 and 3.

Therefore, the wide range of possible alternative distributions selected for the simulation exercise produces a heterogeneous set of patterns for the FSD probabilities. This allows an exhaustive assessment of the performance of the tests described in the previous section under very general contexts and scenarios.

Simulation results

Since the focus of our study is on small samples, for each alternative scenario we consider 4 sample dimensions, i.e. n = 20, 30, 40 and 50. In this section, we present only the results obtained with n = 20. The simulation outcomes for the other sample dimensions (available upon request) are numerically different, but confirm the patterns described for the smallest dimension. Even though the asymptotic critical values for some of the implemented tests are available, the small dimensions of the simulated samples recommends to use exact critical values. We then computed the 1% critical value for each test and each different sample size through one million simulations of Benford distributed values. For each alternative scenario listed in Table 1, we generate 100,000 simulations and calculate the rejection rates of each test.

The three alternative FSD patterns considered reduces to Eq (2) for particular values of its parameter. In these cases, rejection rates measure the empirical size of the tests. Table 2 presents the percentage rejection rates obtained when the FSD patterns corresponds exactly to Eq (2). The values range from 0.901% to 1.064% and confirm a suitable accuracy of the empirical size, not significantly different from the nominal value of 1%.

Table 2. Rejection rates (×100) when the Benford null (3) is true.

Test GB(0) R(-1) H(1) H(2)
χ 2 1.027 0.986 1.034 1.034
W 2 0.979 1.001 1.026 0.969
U 2 0.999 1.041 1.003 0.990
A 2 0.996 0.980 0.998 0.982
KS d 0.995 0.999 1.027 0.989
KU d 0.959 1.010 1.011 0.980
MT 0.922 0.901 0.958 0.907
KS s 0.972 1.064 1.031 0.992
KU s 1.040 1.061 1.034 1.008
AD 0.972 0.973 0.996 0.970
Q 1.001 0.984 1.042 1.045

Power against alternative FSD patterns

Table 3 shows the rejection rates of the tests against the three alternative FSD for a subset of the parameters considered in the Monte Carlo exercise (as before, the full set of results is available upon request). The panel reserved to GB alternatives highlights that, independently of the value of θ, A2, W2, AD, KSd and MT seem to offer the best general performances, whereas the power of KSs is optimal when θ < 0, but deteriorates when θ > 0. All the other tests seem to offer rejection rates always below average. When the digits are uniformly distributed (i.e. θ = 1), the power of MT, W2 and U2 is slightly less than 60%, and the one of AD and KSd is close to 50%. All the other tests considered exhibit a power between 30 and 35%.

Table 3. Rejection rates when the FSD is distribuited according to Eqs (13), (14) and (15).

Generalized Benford
θ -1.5 -1 -0.5 0.5 1 1.5
χ 2 0.224 0.039 0.004 0.079 0.336 0.713
W 2 0.907 0.543 0.115 0.127 0.573 0.924
U 2 0.776 0.366 0.066 0.055 0.300 0.707
A 2 0.896 0.519 0.103 0.137 0.593 0.931
KS d 0.848 0.463 0.094 0.110 0.498 0.874
KU d 0.690 0.287 0.048 0.072 0.380 0.788
MT 0.826 0.413 0.070 0.145 0.592 0.926
KS s 0.888 0.538 0.131 0.052 0.339 0.761
KU s 0.689 0.304 0.060 0.050 0.283 0.677
AD 0.832 0.430 0.080 0.112 0.534 0.906
Q 0.280 0.069 0.010 0.069 0.301 0.670
Rodriguez
β -15 -11 -5 1 5 9
χ 2 0.227 0.197 0.100 0.041 0.155 0.215
W 2 0.353 0.286 0.096 0.044 0.281 0.394
U 2 0.171 0.140 0.066 0.075 0.219 0.256
A 2 0.384 0.321 0.124 0.037 0.259 0.378
KS d 0.325 0.272 0.110 0.045 0.249 0.343
KU d 0.230 0.188 0.077 0.068 0.227 0.287
MT 0.413 0.356 0.152 0.021 0.189 0.322
KS s 0.196 0.156 0.053 0.018 0.140 0.212
KU s 0.169 0.139 0.065 0.054 0.179 0.224
AD 0.346 0.290 0.115 0.021 0.187 0.299
Q 0.203 0.178 0.092 0.035 0.133 0.188
Hürlimann
ρ 0.1 0.3 0.6 4 7 10
χ 2 0.966 0.438 0.045 0.069 0.423 0.779
W 2 0.387 0.079 0.020 0.038 0.140 0.274
U 2 0.977 0.392 0.035 0.125 0.781 0.987
A 2 0.863 0.232 0.031 0.054 0.273 0.587
KS d 0.751 0.192 0.024 0.058 0.285 0.508
KU d 0.978 0.379 0.034 0.104 0.692 0.965
MT 0.390 0.149 0.034 0.052 0.166 0.266
KS s 0.455 0.085 0.018 0.043 0.187 0.341
KU s 0.940 0.283 0.026 0.088 0.622 0.936
AD 0.844 0.254 0.032 0.057 0.294 0.579
Q 0.967 0.437 0.045 0.073 0.429 0.774

The central panel of Table 3 presents the rejection rates obtained with FSD simulated according to Eq (14). A2, W2, AD, KSd and MT show again the best general properties independently of the value of the parameter, while the power of the remaining tests results smaller. The only exception is the case β = 1, where the top performers are U2 and KUd.

Finally, the output against Hürlimann distributed digits is in the bottom panel of Table 3. In this case, the top performers are U2, KUd and KUs, whereas the power of χ2 and Q, is optimal only when ρ < 1. Instead, the rejection rates of MT, KSs and W2 are always below average.

Power against continuous random variables

Table 4 presents the results obtained with data simulated according to a lognormal distribution with the five selected scale parameters. As expected, the rejection rates are inversely related to the shape parameter. When σ decreases the rejection rates are close to 1, while when the shape approaches to 1, the rejection rates approach to the nominal level of the test. Even though every panel seems to tell a slightly different story, there are some testing procedures that seem to regularly offer desirable properties. In particular, they are KUs, U2 and KUd. Also KSs and KSd yield adequate results on average. On the other side of the ranking, what stands out are the poor performances offered by W2 when μ = 0; by MT when μ = 0, 1 and 1.5; by AD when μ = 1; and by χ2 and Q when μ = 0.5.

Table 4. Rejection rates for Lognormal distributed values.

μ 0 0 0 0.5 0.5 0.5
σ 0.3 0.5 0.7 0.3 0.5 0.7
χ 2 0.669 0.112 0.024 0.468 0.021 0.007
W 2 0.219 0.057 0.019 0.985 0.337 0.044
U 2 0.984 0.271 0.031 0.991 0.313 0.037
A 2 0.428 0.078 0.023 0.976 0.296 0.037
KS d 0.439 0.101 0.023 0.991 0.356 0.044
KU d 0.957 0.216 0.028 0.986 0.278 0.031
MT 0.207 0.067 0.023 0.855 0.172 0.022
KS s 0.502 0.107 0.025 0.996 0.434 0.064
KU s 0.993 0.309 0.035 0.992 0.312 0.036
AD 0.858 0.150 0.030 0.964 0.256 0.034
Q 0.801 0.125 0.024 0.547 0.039 0.009
μ 1 1 1 1.5 1.5 1.5
σ 0.3 0.5 0.7 0.3 0.5 0.7
χ 2 0.623 0.048 0.007 0.859 0.142 0.026
W 2 0.355 0.025 0.009 0.918 0.176 0.023
U 2 0.940 0.212 0.027 0.998 0.397 0.045
A 2 0.342 0.020 0.008 0.889 0.141 0.020
KS d 0.415 0.049 0.012 0.889 0.195 0.025
KU d 0.910 0.201 0.025 0.987 0.314 0.041
MT 0.054 0.025 0.008 0.015 0.033 0.014
KS s 0.654 0.089 0.018 0.855 0.109 0.010
KU s 0.993 0.314 0.035 0.992 0.309 0.035
AD 0.064 0.005 0.005 0.658 0.074 0.013
Q 0.645 0.052 0.008 0.836 0.129 0.023
μ 2 2 2
σ 0.3 0.5 0.7
χ 2 0.852 0.180 0.035
W 2 0.904 0.225 0.035
U 2 0.936 0.182 0.023
A 2 0.909 0.235 0.039
KS d 0.944 0.260 0.039
KU d 0.921 0.215 0.031
MT 0.909 0.256 0.045
KS s 0.901 0.172 0.021
KU s 0.992 0.308 0.036
AD 0.942 0.231 0.036
Q 0.870 0.176 0.033

Table 5 shows instead the rejection rates obtained with Weibull distributed values for fixed scale coefficients. As expected, now the rejection rates are directly related to the shape parameter. Thus, when b increases, the power approaches to 1. Again, the outstanding performers are KUs, U2 and KUd, followed by KSs and KSd that also offer outcomes above the average. Particularly negative results are showed by MT when a = 0.5 and 1.5; and by χ2 and Q when a = 2.

Table 5. Rejection rates for Weibull distributed values.

a 0.5 0.5 0.5 1 1 1
b 2 3 4 2 3 4
χ 2 0.111 0.507 0.891 0.085 0.329 0.615
W 2 0.149 0.488 0.834 0.050 0.183 0.356
U 2 0.279 0.881 0.995 0.141 0.533 0.864
A 2 0.121 0.429 0.804 0.068 0.267 0.514
KS d 0.154 0.484 0.798 0.078 0.294 0.533
KU d 0.233 0.821 0.991 0.118 0.425 0.749
MT 0.038 0.012 0.002 0.075 0.274 0.471
KS s 0.082 0.345 0.687 0.072 0.269 0.531
KU s 0.214 0.821 0.992 0.213 0.822 0.993
AD 0.068 0.230 0.577 0.115 0.534 0.912
Q 0.103 0.481 0.876 0.093 0.441 0.826
a 1.5 1.5 1.5 2 2 2
b 2 3 4 2 3 4
χ 2 0.028 0.296 0.789 0.014 0.113 0.372
W 2 0.176 0.634 0.936 0.191 0.703 0.948
U 2 0.263 0.870 0.997 0.153 0.714 0.969
A 2 0.155 0.610 0.934 0.161 0.640 0.921
KS d 0.183 0.649 0.949 0.239 0.841 0.992
KU d 0.205 0.788 0.989 0.147 0.738 0.981
MT 0.052 0.179 0.412 0.115 0.469 0.783
KS s 0.261 0.780 0.976 0.295 0.872 0.995
KU s 0.213 0.821 0.993 0.212 0.821 0.992
AD 0.156 0.604 0.920 0.117 0.548 0.887
Q 0.040 0.332 0.803 0.024 0.182 0.477
a 2.5 2.5 2.5
b 2 3 4
χ 2 0.024 0.214 0.596
W 2 0.084 0.492 0.869
U 2 0.117 0.606 0.937
A 2 0.068 0.412 0.801
KS d 0.122 0.675 0.957
KU d 0.106 0.557 0.915
MT 0.075 0.377 0.663
KS s 0.179 0.785 0.986
KU s 0.211 0.821 0.992
AD 0.035 0.236 0.618
Q 0.030 0.244 0.649

General comments

The comparison of different BL compliance tests in a simulated environment provided several important conclusions. First of all, it is not necessarily true that in very small samples there is insufficient power to meaningfully detect or confirm conformance with the BL. Simulations showed that a careful choice of the testing procedure allows to significantly increase the expected power. Consider, for example, the case of the uniform alternative. As previously stressed, this is the FSD distribution that we should expect when numbers are falsified by a manipulator that believes that the proportion of ones, twos,…, nines in the first digit should be equal. By choosing the usual χ2 test, as most of the analysts do in practice, we can expect a power of 33.5%. But if we consider A2 or MT the expected power almost doubles.

Also in other scenarios, simulation results confirmed that the power of χ2 is usually dominated by the one offered by alternative testing procedures. Thus, despite its popularity, the use of this test should be avoided, at least in very small samples. The other testing procedures considered in the simulation exercise showed very different characteristics, and there is not one that strictly dominates the others. It is also interesting to note that the statistics aimed to test the simplest null (2) work properly also when the deviation affects all the digits, and are not necessary dominated by their counterparts that consider the complete joint digits distribution. This is, for example, the case of U2, which is optimal not only when the FSD follows a Hürlimann distribution, but also when the simulated values are Weibull distributed with scale parameter equal to 0.5. However, it performs poorly when the FSD is simulated according to a Generalized Benford distribution with θ > 0 and to a Rodriguez distribution with β < −1. This sub-optimal behavior is common also to other testing procedures. Therefore, it is not possible to identify a test that is able to offer suitable expected performances independently of the alternative scenario. One alternative option could be to define a combined test that merges together the positive features offered by the single tests, as described in the following section.

Combining tests for improving the general performance

The simulation results provided an exhaustive of the small sample behavior of different BL conformity tests. In particular, two of them offered the best general results across the different alternative scenarios: KUs and U2. The idea is then to combine the p-values of these tests in order to derive a procedure that offers a desirable level of power in all the alternative scenarios proposed. However, both offered poor performances when the FSD was distributed according to the Generalized Benford and the Rodriguez distribution. Therefore, in the definition of the combined test, we decided to consider also the A2 test, whose performances under the Generalized Benford and the Rodriguez alternatives are optimal. Among the possible choices for combining the p-values, we chose the minimum function (see [33] for more details on this issue). Thus, the statistics is given by:

Γ{KUs,U2,A2}=1-min{π(KUs),π(U2),π(A2)} (16)

where π(⋅) is the p-value function, defined as 1 − F(⋅). As usual, the p-values of the single tests and the 1% critical values of the combined test Γ will be calculated through the one million simulations of Benford distributed numbers. Our expectation is that the performances of Γ represent a favorable compromise of the ones of KUs, U2 and A2.

Figs 24 provide an immediate assessment of the behavior of Γ with respect to the three single tests. In addition, they allow a direct comparison with the maximum and the minimum rejection rates obtained with all the tests considered in the simulation exercise. The first thing to notice is that the combined test is always the second best choice, independently of which of the three single tests was the best choice. Secondly, the power of Γ is always closer to the best among KUs, U2 and A2, than to the worst. These two features allow the combined test to offer convincing outcomes even when one or two single tests perform below the average, Consider, for example, the case represented in PowerCombinedGB for θ > 0: the power of KUs, U2 is very close to the absolute minimum, whereas the rejection rate of Γ is very close to the one of A2, which is the absolute maximum. In conclusion, test Γ provides a convenient trade-off of the behaviors of the three single tests in all the scenarios considered.

Fig 2. Rejection rates of the combined test for alternative FSD.

Fig 2

The gray region represents the area between the maximum and the minimum value of the power obtained in the simulations.

Fig 4. Rejection rates of the combined test for Weibull distributed values.

Fig 4

The gray region represents the area between the maximum and the minimum value of the power obtained in the simulations.

Fig 3. Rejection rates of the combined test for lognormal distributed values.

Fig 3

The gray region represents the area between the maximum and the minimum value of the power obtained in the simulations.

Empirical application on international trade data

In this section, we provide some examples of the application of the BL compliance tests on Customs data. The contrast of fraud in this context is crucial. The protection of the financial interests is indeed a fundamental task of the European Union (EU) administration, and the collection of customs duties on imports represents the principal source of traditional own resources of the EU budget [34]. In 2017, the EU’s revenue from customs duties was more than 20 million, representing almost 15% of EU total revenue (source: https://ec.europa.eu/budget/library/biblio/publications/2018/financial-report_en.pdf). Under-reporting the value of the imports is usually the main strategy pursued to pay less duties or excises, or to evade import restrictions and certain anti-dumping measures [35]. Proper statistical analyses provide support for the identification of anomalous trades that may result from unfair commercial strategies [36, 37]. Several studies have shown that statistical methods based also on the Benford Law can be profitably used in this field, aiming to spot traders with more than 50 transactions that are suspected of data manipulation [13, 20, 38]. The possibility to extend the focus also to economic operators with less imports can significantly increase the feasibility of the anti-fraud analysis, especially when the investigations are limited to particular periods, or to particular groups of products. Just to have an idea, Table 6 shows the distribution of traders according to the number of declared imports during the whole 2014 in a single Member State of the European Union (not revealed for confidentiality issues). Considering that one year is the usual period length of Customs audit, it is evident how extending the focus also to less active importers allows to almost double the number of economic operators we can monitor. In addition, a reliable and accurate testing procedure for small samples allows to tighten the number of suspicious imports to investigate, reducing the costs of the demanding and time consuming controls necessary to prove the eventual fraud.

Table 6. Number of imports per trader in 2014 for an EU member state.

Number of imports per year Number of traders
less than 20 imports 210,521
between 20 and 29 imports 6,627
between 30 and 39 imports 3,962
between 40 and 49 imports 2,714
50 or more imports 14,459

Consider, for instance, the two examples of the application of the BL compliance tests represented in Table 7. Data were collected in the context of a specific operation by a Member State of the EU and, for confidentiality reasons, the two traders will be labeled simply as Trader 1 and Trader 2. Trader 1 collected 58 imports during three years. The p-values of the tests a suggest a potential departure of the values declared from the theoretical distribution. Most of the p-values are indeed very close or even smaller than 1%, with only one of them larger than 5%. A year-by-year analysis reveals that the 21 import values declared in 2015 yield the smallest p-value, and should therefore be the first to investigate in the search for frauds. Actually, one of those 21 imports was checked by Custom authority, and the corresponding import resulted to be under-valuated. The analysis of Trader 2 provides another interesting pattern. Also in this case, the 94 imports were traded in a three years period, and the p-values of all the tests suggest BL compliance of the declared values. However, the year-by-year analysis identifies a significant deviation from the Benford distribution for the 37 values declared in 2014. Actually, 6 of them were checked by Custom authority, and the corresponding imports resulted to be under-valuated.

Table 7. P-values of Benford tests applied on the import values declared by two traders.

Trader 1
Period 2013–2015 2013 2014 2015
n 58 18 19 21
χ 2 0.010 0.025 0.144 0.018
W 2 0.020 0.065 0.697 0.034
U 2 0.000 0.009 0.516 0.003
A 2 0.020 0.079 0.558 0.037
KS d 0.031 0.088 0.406 0.050
KU d 0.001 0.006 0.279 0.007
MT 0.619 0.481 0.546 0.454
KS s 0.019 0.092 0.281 0.061
KU s 0.000 0.006 0.420 0.002
AD 0.041 0.135 0.732 0.119
Q 0.014 0.023 0.176 0.023
Γ{KUs,U2,A2} 0.000 0.013 0.585 0.004
Trader 2
Period 2012–2014 2012 2013 2014
n 94 31 26 37
χ 2 0.735 0.412 0.870 0.013
W 2 0.155 0.125 0.774 0.000
U 2 0.536 0.233 0.894 0.002
A 2 0.145 0.147 0.784 0.000
KS d 0.322 0.241 0.745 0.001
KU d 0.574 0.350 0.935 0.002
MT 0.111 0.177 0.455 0.002
KS s 0.225 0.273 0.484 0.001
KU s 0.617 0.398 0.780 0.011
AD 0.231 0.164 0.878 0.001
Q 0.758 0.457 0.925 0.018
Γ{KUs,U2,A2} 0.247 0.248 0.898 0.001

In conclusion, the availability of suitable and powerful procedures for testing BL compliance in small samples can efficiently support the Custom anti-fraud activities. It allows not only to extend the range of applicability through the possibility of use them for traders with few operations, but also to identify small subsets of imports that deserve further investigations.

Conclusion

The popularity of the BL is remarkably increasing in these last years. Since a set of numbers is expected to be Benford compliant under very general conditions, it can be used in many applications where the target is to identify potentially manipulated figures, as, for example, anti-fraud investigations. The automatic assessment of BL conformity through statistical methods requires testing procedures with desirable statistical properties. Especially in audit where many samples are tested, we aim for a statistical test which controls the number of false alarms and guarantees at the same time a suitable level of power. The choice of the Benford compliance test is even more important when the number of observations in the sample is small, given that suitable level of powers are very difficult to achieve.

The aim of this article was first of all to provide an extensive analysis of the performances of several BL compliance tests in very small samples. Simulation results proved that small sample properties can significantly vary, depending on the alternative scenario. In general, every test alternates good and average results across all the alternatives and there is no a procedure that strictly dominates the others. However, the following regularities emerged from the simulation exercise:

  • despite its popularity, the power of Pearson’s χ2 test was often below average;

  • KUs and U2 achieved outstanding results in most of the alternative scenarios considered, but not when the FSD is uniformly distributed.

The need of a procedure with optimal results independently of the alternative digits pattern encouraged us to define a new test. The main idea was to merge the positive small sample properties offered by some of the tests in the simulation experiment. Therefore we proposed to combines the p-values of KUs, U2 and A2 through the min function. The resulting test achieved always desirable levels of powers in all alternative designs, even when one or two single tests performed below the average.

Finally, an empirical application presented a practical case where the availability of a reliable testing procedure for small samples allows (i) to increase the number of samples under investigation; and (ii) a more accurate selection of the cases that are suspected of data manipulation. This second issue is particularly relevant in anti-fraud audits. Non-conformity to the BL does not necessarily mean that the corresponding economic operator is a fraudster. Further, sometimes expensive, investigations are required to prove the irregularities. Therefore, having the possibility to focus the attention on a restricted number of suspect cases together with a reliable control over the number of false alarms are two essential requirements for an affordable anti-fraud analysis.

Supporting information

S1 File

(ZIP)

Data Availability

All relevant data are within the paper and its Supporting information files.

Funding Statement

The author received no specific funding for this work.

References

  • 1. Newcomb S. Note on the frequency of use of the different digits in natural numbers. American Journal of Mathematics. 1881;4(1):39–40. doi: 10.2307/2369148 [DOI] [Google Scholar]
  • 2. Benford F. The law of anomalous numbers. Proceedings of the American Philosophical Society. 1938; p. 551–572. [Google Scholar]
  • 3. Hill TP. A statistical derivation of the significant-digit law. Statistical Science. 1995;10(4):354–363. doi: 10.1214/ss/1177009869 [DOI] [Google Scholar]
  • 4. Berger A, Hill TP. A basic theory of Benford’s Law. Probability Surveys. 2011;8:1–126. doi: 10.1214/11-PS175 [DOI] [Google Scholar]
  • 5. Berger A, Hill TP. An introduction to Benford’s law. Princeton University Press; 2015. [Google Scholar]
  • 6. Miller SJ. Benford’s Law. Princeton University Press; 2015. [Google Scholar]
  • 7. Ley E. On the peculiar distribution of the US stock indexes’ digits. The American Statistician. 1996;50(4):311–313. doi: 10.1080/00031305.1996.10473558 [DOI] [Google Scholar]
  • 8. Nigrini MJ, Miller SJ. Benford’s law applied to hydrology data—results and relevance to other geophysical data. Mathematical Geology. 2007;39(5):469–490. doi: 10.1007/s11004-007-9109-5 [DOI] [Google Scholar]
  • 9. Geyer A, Martí J. Applying Benford’s law to volcanology. Geology. 2012;40(4):327–330. doi: 10.1130/G32787.1 [DOI] [Google Scholar]
  • 10. Kennedy AP, Yam SCP. On the authenticity of COVID-19 case figures. PloS One. 2020;15(12):e0243123. doi: 10.1371/journal.pone.0243123 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11. Diekmann A. Not the first digit! using benford’s law to detect fraudulent scientif ic data. Journal of Applied Statistics. 2007;34(3):321–329. doi: 10.1080/02664760601004940 [DOI] [Google Scholar]
  • 12. Golbeck J. Benford’s law applies to online social networks. PloS One. 2015;10(8):e0135169. doi: 10.1371/journal.pone.0135169 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13. Cerioli A, Barabesi L, Cerasa A, Menegatti M, Perrotta D. Newcomb–Benford law and the detection of frauds in international trade. Proceedings of the National Academy of Sciences. 2019;116(1):106–115. doi: 10.1073/pnas.1806617115 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14. Durtschi C, Hillison W, Pacini C. The effective use of Benford’s law to assist in detecting fraud in accounting data. Journal of Forensic Accounting. 2004;5(1):17–34. [Google Scholar]
  • 15.Goodman WM. Reality checks for a distributional assumption: The case of “Benford’s Law”. In: Joint Statistical Meeting–Business and Economic Statistics Section. Citeseer; 2013. p. 2789–2803.
  • 16. Günnel S, Tödter KH. Does Benford’s Law hold in economic research and forecasting? Empirica. 2009;36(3):273–292. doi: 10.1007/s10663-008-9084-1 [DOI] [Google Scholar]
  • 17. Geyer CL, Williamson PP. Detecting fraud in data sets using Benford’s Law. Communications in Statistics-Simulation and Computation. 2004;33(1):229–246. doi: 10.1081/SAC-120028442 [DOI] [Google Scholar]
  • 18.Joenssen D. Testing for benford’s law: A monte carlo comparison of methods. Available at SSRN 2545243. 2014.
  • 19. Lesperance M, Reed W, Stephens M, Tsao C, Wilton B. Assessing conformance with Benford’s Law: Goodness-of-fit tests and simultaneous confidence intervals. PloS One. 2016;11(3):e0151235. doi: 10.1371/journal.pone.0151235 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20. Barabesi L, Cerasa A, Cerioli A, Perrotta D. On Characterizations and Tests of Benford’s Law. Journal of the American Statistical Association. 2021;0(0):1–17. doi: 10.1080/01621459.2021.1891927 [DOI] [Google Scholar]
  • 21. Morrow J. Benford’s Law, families of distributions and a test basis. Centre for Economic Performance. 2014. [Google Scholar]
  • 22. Choulakian V, Lockhart RA, Stephens MA. Cramér-von Mises statistics for discrete distributions. The Canadian Journal of Statistics/La Revue Canadienne de Statistique. 1994; p. 125–137. doi: 10.2307/3315828 [DOI] [Google Scholar]
  • 23. Lockhart RA, Spinelli JJ, Stephens MA. Cramér–von Mises statistics for discrete distributions with unknown parameters. The Canadian Journal of Statistics/La Revue Canadienne de Statistique. 2007; p. 125–133. doi: 10.1002/cjs.5550350111 [DOI] [Google Scholar]
  • 24.Kuiper NH. Tests concerning random points on a circle. In: Nederl. Akad. Wetensch. Proc. Ser. A. vol. 63; 1960. p. 38–47.
  • 25. Diaconis P, et al. The distribution of leading digits and uniform distribution mod 1. The Annals of Probability. 1977;5(1):72–81. doi: 10.1214/aop/1176995891 [DOI] [Google Scholar]
  • 26. Kazemitabar J., Kazemitabar J. Benford Test Based on Logarithmic Property. International Journal of Auditing Technology. doi: 10.1504/IJAUDIT.2022.10047587 [DOI] [Google Scholar]
  • 27. Anderson TW, Darling DA. A test of goodness of fit. Journal of the American Statistical Association. 1954;49(268):765–769. doi: 10.1080/01621459.1954.10501232 [DOI] [Google Scholar]
  • 28. Barabesi L, Pratelli L. On the Generalized Benford law. Statistics & Probability Letters. 2020;160:108702. doi: 10.1016/j.spl.2020.108702 [DOI] [Google Scholar]
  • 29. Pietronero L, Tosatti E, Tosatti V, Vespignani A. Explaining the uneven distribution of numbers in nature: the laws of Benford and Zipf. Physica A: Statistical Mechanics and its Applications. 2001;293(1-2):297–304. doi: 10.1016/S0378-4371(00)00633-6 [DOI] [Google Scholar]
  • 30. Rodriguez RJ. First significant digit patterns from mixtures of uniform distributions. The American Statistician. 2004;58(1):64–71. doi: 10.1198/0003130042782 [DOI] [Google Scholar]
  • 31. Hürlimann W. Generalizing Benford’s law using power laws: application to integer sequences. International journal of Mathematics and Mathematical Sciences. 2009;2009. [Google Scholar]
  • 32. Rodriguez RJ. Reducing false alarms in the detection of human influence on data. Journal of Accounting, Auditing & Finance. 2004;19(2):141–158. doi: 10.1177/0148558X0401900202 [DOI] [Google Scholar]
  • 33. Heard NA, Rubin-Delanchy P. Choosing between methods of combining p-values. Biometrika. 2018;105(1):239–246. doi: 10.1093/biomet/asx076 [DOI] [Google Scholar]
  • 34. Erkoreka M. The European Union Customs Administration and the Fight Against Fraud. European Papers-A Journal on Law and Integration. 2021;2020(3):1425–1434. [Google Scholar]
  • 35.European Commission. Operation SNAKE: EU and Chinese customs join forces to target undervaluation of goods at customs. Press release IP-14-1001 Available at europaeu/rapid/. 2014.
  • 36. Cerasa A, Cerioli A. Outlier-free merging of homogeneous groups of pre-classified observations under contamination. Journal of Statistical Computation and Simulation. 2017;87(15):2997–3020. doi: 10.1080/00949655.2017.1351564 [DOI] [Google Scholar]
  • 37. Perrotta D, Torti F. Detecting price outliers in European trade data with the forward search. In: Data Analysis and Classification. Springer; 2010. p. 415–423. [Google Scholar]
  • 38. Barabesi L, Cerasa A, Cerioli A, Perrotta D. Goodness-of-fit testing for the Newcomb-Benford law with application to the detection of customs fraud. Journal of Business & Economic Statistics. 2018;36(2):346–358. doi: 10.1080/07350015.2016.1172014 [DOI] [Google Scholar]

Decision Letter 0

Miguel A Fernández

1 Jun 2022

PONE-D-21-17178Testing for Benford's Law in very small samples: simulation study and a new test proposalPLOS ONE

Dear Dr. Cerasa,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process. I have received two reports for this paper. Both reviewers seem to be happy with the paper. They have made only minor suggestions for the paper. Please handle them carefully.

Please submit your revised manuscript by Jul 16 2022 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Miguel A. Fernández, Ph.D.

Academic Editor

PLOS ONE

Journal Requirements:

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at 

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and 

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. We note that you have stated that you will provide repository information for your data at acceptance. Should your manuscript be accepted for publication, we will hold it until you provide the relevant accession numbers or DOIs necessary to access your data. If you wish to make changes to your Data Availability statement, please describe these changes in your cover letter and we will update your Data Availability statement to reflect the information you provide.

3. Please include captions for your Supporting Information files at the end of your manuscript, and update any in-text citations to match accordingly. Please see our Supporting Information guidelines for more information: http://journals.plos.org/plosone/s/supporting-information.

4. Please review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript. If you need to cite a retracted article, indicate the article’s retracted status in the References list and also include a citation and full reference for the retraction notice.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Yes

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: Yes

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: Introducing new Benford tests that work for small and/or large datasets is welcome as the existing tests usually do not meet the expectations. In this regards, I recommend publication of this work. However, one aspect of any good research is exploring and citing relevant work. For example, one work I am aware of which has not been cited in the paper under review is the following:

Benford Test Based on Logarithmic Property

by Javad Kazemitabar, Jalil Kazemitabar

International Journal of Auditing Technology

DOI: 10.1504/IJAUDIT.2022.10047587

As a reviewer, I require this change to be made.

Reviewer #2: A few minor suggestions/corrections: p 10 line 198 and figures 1 and 2, umlaut ü on Hürlimann; p 12 line 221 "not necessarily"; p 16 line 282 "almost 15% of the EU"; references 3 and 4 have no "et al"; references 1,2,14,26,30, 30 are missing capitals on journal titles; reference 32 should read "combining p-values".

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2022 Jul 22;17(7):e0271969. doi: 10.1371/journal.pone.0271969.r002

Author response to Decision Letter 0


30 Jun 2022

Reviewer #1: Introducing new Benford tests that work for small and/or large datasets is welcome as the existing tests usually do not meet the expectations. In this regards, I recommend publication of this work. However, one aspect of any good research is exploring and citing relevant work. For example, one work I am aware of which has not been cited in the paper under review is the following:

Benford Test Based on Logarithmic Property by Javad Kazemitabar, Jalil Kazemitabar

International Journal of Auditing Technology, DOI: 10.1504/IJAUDIT.2022.10047587

As a reviewer, I require this change to be made.

Thanks a lot for your interesting bibliographic suggestion. It was added to the revised version of the manuscript.

Reviewer #2: A few minor suggestions/corrections: p 10 line 198 and figures 1 and 2, umlaut ü on Hürlimann; p 12 line 221 "not necessarily"; p 16 line 282 "almost 15% of the EU"; references 3 and 4 have no "et al"; references 1,2,14,26,30, 30 are missing capitals on journal titles; reference 32 should read "combining p-values".

Thanks a lot for spotting these mistakes. They are all fixed in the new version of the manuscript.

Attachment

Submitted filename: ReplyReviewer2.docx

Decision Letter 1

Miguel A Fernández

12 Jul 2022

Testing for Benford's Law in very small samples: simulation study and a new test proposal

PONE-D-21-17178R1

Dear Dr. Cerasa,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Miguel A. Fernández, Ph.D.

Academic Editor

PLOS ONE

Additional Editor Comments (optional):

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #1: All comments have been addressed

Reviewer #2: All comments have been addressed

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: All my concerns are addressed. The specific topic covered in this paper is important in the sense that many of the existing Benford tests are not general enough. In other words, they work in some scenarios such as medium size dataset, but fail to cover small or large datasets. As such I welcome the manuscript under review.

Reviewer #2: (No Response)

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: Yes: Javad Kazemitabar

Reviewer #2: No

**********

Acceptance letter

Miguel A Fernández

14 Jul 2022

PONE-D-21-17178R1

Testing for Benford’s Law in very small samples: simulation study and a new test proposal

Dear Dr. Cerasa:

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department.

If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org.

If we can help with anything else, please email us at plosone@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr Miguel A. Fernández

Academic Editor

PLOS ONE


Articles from PLoS ONE are provided here courtesy of PLOS

RESOURCES