We read with interest the paper by Wang et al. (2024) on confusion in public goods games (1). Their conclusion—that confusion is not a significant factor in explaining cooperation— challenges much of the existing literature (2–5). However, we believe that their study overlooks crucial aspects of confusion measurement and lacks critical design elements.
Wang et al. suggest that the use of 10 control questions with feedback explains the low levels of confusion they find. They argue that the confusion observed in previous studies is due to insufficient instruction. However, this conclusion is drawn without comparing their design to a baseline treatment similar to earlier studies with fewer control questions and less feedback. Such a comparison is essential to substantiate their claim. Without it, the interpretation of their findings remains speculative.
Moreover, Wang et al. also test confusion by comparing public good contributions in interaction between human group members with interaction with computerized group members—an often-used method (2). Wang et al. observe substantial cooperation even in their computer treatment, which seems to contradict their assertion that confusion is not an issue. This finding itself raises questions about the validity of their claim.
Furthermore, their study does not engage with earlier literature using similar but more comprehensive multiitem ways of measuring confusion. For example, Fosgaard et al. (5) test participants’ understanding of both sides of the social dilemma through an incentivized task. Their six-item task captures the understanding of the payoff-maximizing strategy as well as the understanding of the socially optimal strategy. Such tests are better suited for ascertaining whether participants grasp the social dilemma of engaging in public good contribution and, if not, in which direction participants’ errors occur.
Results from Fosgaard et al. (5), shown in Fig. 1, highlight that substantial confusion is detected with the more comprehensive measures and that confusion exists on both sides of the social dilemma. Moreover, varying the way the public good game is framed—Give vs. Take framing—substantially affects the measured degree of confusion. Ferraro and Vossler (3) also indicate this finding of treatment-dependent confusion, reporting that labeling contributions as donations rather than investments leads to lower contribution levels when other group members are computers. Given this, making bold conclusions about the importance of confusion based on one treatment without proper baselines, as Wang et al. (1), appears premature.
In conclusion, while Wang et al. provide intriguing evidence of confusion in public goods games, their findings risk underestimating its role in shaping cooperation and responses to experimental treatments.
Acknowledgments
Author contributions
T.R.F. and E.W. wrote the paper.
Competing interests
The authors declare no competing interest.
References
- 1.Wang G., et al. , Confusion cannot explain cooperative behavior in public goods games. Proc. Natl. Acad. Sci. U.S.A. 121, e2310109121 (2024), 10.1073/pnas.2310109121. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.Houser D., Kurzban R., Revisiting kindness and confusion in public goods experiments. Am. Econ. Rev. 92, 1062–1069 (2002). [Google Scholar]
- 3.Ferraro P. J., Vossler C. A., The source and significance of confusion in public goods experiments. B.E. J. Econ. Anal. Policy 10, 53 (2010). [Google Scholar]
- 4.Burton-Chellew M. N., El Mouden C., West S. A., Conditional cooperation and confusion in public-goods experiments. Proc. Natl. Acad. Sci. U.S.A. 113, 1291–1296 (2016). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5.Fosgaard T. R., Hansen L. G., Wengström E., Framing and misperception in public good experiments. Scand. J. Econ. 119, 435–456 (2017). [Google Scholar]