Skip to main content
Entropy logoLink to Entropy
. 2019 Mar 21;21(3):307. doi: 10.3390/e21030307

Correction: Veen, D.; Stoel, D.; Schalken, N.; Mulder, K.; Van de Schoot, R. Using the Data Agreement Criterion to Rank Experts’ Beliefs. Entropy 2018, 20, 592

Duco Veen 1,*, Diederick Stoel 2, Naomi Schalken 1, Kees Mulder 1, Rens van de Schoot 1,3
PMCID: PMC7514787  PMID: 33267530

Abstract

Due to a coding error the marginal likelihoods have not been correctly calculated for the empirical example and thus the Bayes Factors following from these marginal likelihoods are incorrect. The corrections required occur in Section 3.2 and in two paragraphs of the discussion in which the results are referred to. The corrections have limited consequences for the paper and the main conclusions hold. Additionally typos in Equations, and, an error in the numbering of the Equations are remedied.

1. Corrections in Section 3.2

  • Corrected Table 2
    KL Divergence DACd DACd Ranking md(y) & mJ(y) BFJd BFJd Ranking
    Expert 1 1.43 0.56 2 5.57 × 10−68 0.21 3
    Expert 2 2.86 1.12 3 6.82 × 10−68 0.17 2
    Expert 3 5.76 2.26 4 2.19 × 10−69 5.31 4
    Expert 4 0.19 0.07 1 1.72 × 10−67 0.07 1
    Benchmark 2.55 - - 1.16 × 10−68 - -
  • Corrected Table 3
    U(0,5) U(−10,10) N(0,102) N(0,103) N(0,104)
    KL[πJ(.|y)||π1] 1.43 1.42 1.37 1.42 1.42
    KL[πJ(.|y)||π2] 2.86 2.84 2.75 2.85 2.85
    KL[πJ(.|y)||π3] 5.76 5.75 5.67 5.76 5.77
    KL[πJ(.|y)||π4] 0.19 0.19 0.20 0.19 0.19
    KL[πJ(.|y)||πJ] 2.55 3.93 4.18 6.46 8.76
    mJ(y) 1.16 × 10−68 2.91 × 10−69 5.65 × 10−69 2.26 × 10−69 7.33 × 10−70
  • Corrected Table 4
    Expert 1 Expert 2 Expert 3 Expert 4
    KL Ratio BF KL Ratio BF KL Ratio BF KL Ratio BF
    Expert 1 1 1 0.50 0.82 0.25 25.42 7.63 0.32
    Expert 2 2.00 1.22 1 1 0.50 31.13 15.23 0.40
    Expert 3 4.03 0.04 2.02 0.03 1 1 30.75 0.01
    Expert 4 0.13 3.09 0.07 2.52 0.03 78.54 1 1

2. New Version of Section 3.2 Paragraphs 3–6

The results of Table 2 show that expert four provided the best prediction out of the experts, when using both the DACd and the BFJd. Experts one and two provided similar predictions concerning their tacit knowledge; they expected almost the same value for the location parameter; however, expert one was less certain about this prediction (see Table 1). As the prediction of the location was not entirely correct, the increased uncertainty of expert one means that this expert provided more plausibility to the regions of the parameter space that were also supported by the data. Here we see the difference between DACd and the BFJd arise as discussed in Section 2.3. Overconfidence ifs penalized more severely by the DACd and as such the conclusion on which expert would be preferred changes between experts one and two depending on which measure you use. When we look at the DACd, in the case when πJ(θ) is the U(0,5) density, the additional penalization of the overconfidence even causes a different conclusion between experts one and two, namely, expert one is in prior-data agreement and expert two is in prior-data disagreement. For the BFJd both are concluded to be in agreement with the data. Expert three provided a prediction that, to a large extent, did not support the same parameter space as the data. In fact, expert three provides a lot of support for regions of the parameter space that the data did not support. The discrepancy between expert three and the data was of such proportions that, besides expert two, we also concluded a prior-data disagreement to exist for expert three. If we had no information beforehand, except knowing the region within which the average turnover per professional could fall, we would have lost less information than by considering the predictions of experts two and three. The BFJd differs from the DACd in the sense that when πJ(θ) is the U(0,5) density, the benchmark only outperforms expert 3.

From the sensitivity analyses of Table 3 we can find that the reference posterior remains quite stable and therefore the KL divergences for the experts do not change substantially; however, the changing KL divergence for the benchmark would shift the prior-data disagreement boundary. When πJ(θ) was the N(0,103) or N(0,104) density, expert three would no longer be in prior-data conflict, whilst prior-data disagreement for expert two was only concluded if πJ(θ) was the U(0,5) density. For the BF changing the benchmark also shifts the prior-data (dis)agreement boundary arbitrarily. In this case our decisions on prior-data (dis)agreement would only change for the N(0,104) prior, where expert 4 would no longer be in prior-data disagreement. The sensitivity analysis showed that decisions on prior-data (dis)agreement might not be entirely reliable, whilst the ranking of experts remained stable.

Table 4 shows the results when we only compare experts on their KL divergences and their marginal likelihoods and we omit the benchmarks. We see the difference between the BF and the KL divergence ratios when we compare experts one and two. The differences arise from the more severe penalization of overconfidence by KL divergences compared to BF, as discussed in Section 2.3. Using KL divergence ratios we concluded that expert two had twice the amount of loss of information, whilst the BF even favors expert two over expert one with odds of 1.22.

The results of the empirical study show a slight difference in the conclusions with regard to the ranking of the experts depending on which measure we used, DACd or BFJd. Both measures select the same expert as being the best. If decisions should be made concerning average turnover per professional, decision makers would be wise to consult expert four, as this expert seemed to have the best knowledge of the underlying factors driving these results.

3. New Version of Section 4 Paragraph 1

In this paper, we use both the BF and the DAC to rank experts’ beliefs when they are specified in the probabilistic form of prior distributions. When comparing the BF and the DAC, the limiting case example of Section 2.3 springs to mind. In the introduction, we stated that forecasting without specifying uncertainty would not make sense to us and, in that light, we would prefer to use a measure that would classify doing so as undesirable behavior and punish this extreme case. An example of this behavior can be seen in the empirical example where while using the BF we would favor expert two over expert one, however whilst using KL divergences, we would favor expert one over expert two.

4. Nev Version of Section 4 Paragraph 3

One of the reasons for the sensitivity of the DAC to different choices for πJ(θ) can be seen by comparing the KL divergences of expert one and two of the empirical example. As a referee pointed out to us, KL divergences are tail sensitive and this can be seen in this comparison. Expert one is a little more uncertain and as such the tail of π1(θ) overlaps somewhat more with πJ(θ|y) than the tails of π2(θ). This leads to half the loss of information. One could deem this tail sensitivity to be undesirable and, with differently shaped prior distributions, this problem might become more pronounced. If it is deemed undesirable, one could favor using the BF, which actually favors expert two with odds of 1.22 over expert 1. Alternatively, an interesting area for future research could be to investigate the use of alternative divergence measures. A good starting point for finding alternative measures can be found in the Encyclopedia of Distances by Deza and Deza [46].

5. Corrections in Equations

In Section 2.3 [1] only there is a consistent mistake in the brackets involved in the KL functions. There is one “]” bracket too many. The new Equations read:

DAC2, dJ=mJ(y)md(y)exp{KL[πJ(·|y)||πd(·|y)]}= BFJdexp{KL[πJ(·|y)||πd(·|y)]},

exp{KL[πJ(·|y)||πd(·|y)]} and KL[πJ(·|y)||δθ0(·)].

The second change in the Equations is in the numbering. The equations go from (3)–(5), (4) is skipped in the numbering. This made the numbering of the equations flawed after Section 2.2.1. This has been corrected.

Reference

  • 1.Veen D., Stoel D., Schalken N., Mulder K., Van de Schoot R. Using the Data Agreement Criterion to Rank Experts’ Beliefs. Entropy. 2018;20:592. doi: 10.3390/e20080592. [DOI] [PMC free article] [PubMed] [Google Scholar]

Articles from Entropy are provided here courtesy of Multidisciplinary Digital Publishing Institute (MDPI)

RESOURCES