Abstract
How humans and animals distribute their behavior across choice options has been of key interest to economics, psychology, ecology, and related fields. Neoclassical and behavioral economics have provided prescriptions for how decision-makers can maximize their reward or utility, but these formalisms are used by decision-makers rarely. Instead, individuals allocate their behavior in proportion to the worth of their options, a phenomenon captured by the generalized matching law. Why biological decision-makers adopt this strategy has been unclear. To provide insight into this issue, this article evaluates the performance of matching across a broad spectrum of decision situations, using simulations. Matching is found to attain a high or near-optimal gain, and the strategy achieves this level of performance following a single evaluation of the decision options. Thus, matching provides highly efficient decisions across a wide range of choice environments. This result offers a quantitative explanation for the broad adoption of matching by biological decision-makers.
Introduction
How humans and animals allocate behavior, time, and effort to their options has been a fundamental question in science and human society. Neoclassical economics provides a normative apparatus that shows how humans and animals can allocate their time, effort, or monetary resources [1, 2]. A hallmark of this normative set of theories is that decision-makers maximize a criterion—individuals allocate their resources to maximize their total utility or payoff [3–5]. Despite its normative appeal, this framework is adopted by decision-makers infrequently [6, 7].
Biological sciences have reported a much more common strategy in which decision-makers allocate their behavior in proportion to the value of their options [8–13]. This behavior has been observed so commonly that it has been formalized as the matching law [8–10, 12, 14, 15]. When generalized—allowing for the representation of value to be nonlinear [16–23]—the matching law explains much of choice behavior, including undermatching, overmatching, and their specific forms [15, 19, 24–26].
Why decision-makers adopt matching so frequently has been a conundrum to decision sciences, in particular because matching is generally suboptimal [2, 15, 27–29]. Decades-long debates have centered on whether matching might represent a universal heuristic to optimal choice behavior [13, 15, 30–32]. This issue has been difficult to resolve because it is not known how matching performs at a general level and what its specific strength is.
Previous studies assessed the performance of matching in specific tasks, particularly focusing on variable-interval and variable-ratio schedules [8–10, 12, 14, 15]. An analytical study attempted to evaluate the performance of matching at a general level [13], but the analytical approach can only identify the necessary—and not the sufficient—contingencies between utility and effort for which matching is optimal.
To address this issue, this article uses a computational approach to evaluate the performance of matching at a general level. Using this approach, matching is also assessed along decision time, i.e., the necessary number of evaluations of decision options. This methodology uncovers a striking finding: Matching harvests on average around 90% of the total available reward following a single evaluation of the worth of decision options.
Results
Assessment of matching efficiency
This article assesses the performance of matching over a broad spectrum of decision situations. The evaluations include 2, 3, and up to 10 decision options, each with distinct contingency between relative utility and relative effort (Materials and Methods). These contingencies include reinforcement schedules that have been used to assess the performance of matching in previous studies [8–10, 12–15], and generalize the spectrum to many other contingencies and number of options, for a total of 9000 decision settings (Materials and Methods). In each case, the performance of matching is quantified relatively to maximization, which provides the highest possible gain. The performance of matching and maximization is further assessed over the number of evaluations of the decision-makers’ options.
Figure 1 shows the principal finding. Matching decision-makers (green) obtained an average 90.2% of the total possible gain across all choice situations. Notably, they collected this level of gain following a single evaluation of the worth of each option. In contrast, maximizing decision-makers required an average of at least 16 evaluations to reach equivalent performance (Figure 1, gray).
To exemplify this finding, consider a foraging animal that must decide how to allocate her presence across several patches of food. Matching allows her to visit each patch just once to evaluate its worth, and allocate her presence across the patches in proportion to their relative worth. If she maximized reward, she would have to visit each patch at least 16 times to converge on a behavioral allocation of quality comparable to matching.
Thus, matching allows individuals to make effective decisions in a single step, i.e., matching provides highly efficient decisions.
Efficiency of generalized matching
Matching can explain much of choice behavior when generalized [16–20]. The following analysis investigates the efficiency of generalized matching, i.e., performance following a single evaluation. In generalized matching, the relative utilities of choice options are parameterized using an exponent (Materials and Methods). This exponent implements tendencies for undermatching or overmatching . Specifically, decision-makers who undermatch the relative utility proportions are relatively indifferent to reward. Decision-makers who are drawn to rewarding options overmatch the relative utility proportions. In the extreme, highly over-matching, greedy decision-makers put all their effort into the seemingly richest option .
Figure 2A shows the performance of generalized matching as a function of across the space of choice situations. The figure reveals that , which characterizes strict matching [33–35], lies near the optimum, with and delivering a 90.2% and 90.5% gain, respectively. Overmatching and undermatching are similarly effective in their typical, mild forms [25, 26], but incur a performance loss when extreme. In the extremes, the greedy strategy, compared with indifference, provides a relatively high gain over an appreciable range of values (Figure 2A). The following analyses therefore compare the performance of matching also with this degenerate form.
Efficiency across choice situations
In changing environments, decision-makers may apply maximization or matching repeatedly [36–40]. For matching, this repeated application, i.e., , is an instance of melioration [41]. Melioration is shown to increase the performance of matching further. Figure 2B, left green bars, shows that a second evaluation and effort allocation increases the average performance from 90.2% to 92.0%. The performance saturates in about the third and fourth evaluation, reaching 92.3% of the possible maximum. The performance of the greedy strategy does not, by definition, improve with additional evaluations (Figure 2B, left orange bars; see also Figure 3B for a specific example).
Thus far, decision-makers were assumed to allocate effort equally when sampling the initial worth of their options (Materials and Methods). This case also incorporates situations in which individuals can estimate the worth of their options by direct means (e.g., price), in which no initial sampling is required. However, in many cases in nature, humans and animals must sample their options to learn about their initial worth. Equal sampling may be impractical or impossible. The following analysis reveals that the efficiency of matching is maintained regardless of the type of initial distribution of effort. Specifically, Figure 2B, right green bars, shows that decision-makers who distribute their initial effort entirely randomly (e.g., Figure 3B) can still apply the same strategy with comparable level of efficiency. In addition, the performance of matching increases even more steeply by re-adjusting effort in response to additional evaluations . On the fourth evaluation, the performance reaches 91.6%. A two-way ANOVA with factors evaluation number and initial effort distribution detected a significant effect of , a significant effect of , and a significant interaction between the factors . The greedy strategy does not depend on the number of evaluations (Figure 2B, orange) but is sensitive to the initial allocation of effort
Matching maintains its efficiency also in regard to the number of decision options. Figure 2C shows that matching is efficient even when faced with a relatively high number of options. Following a single evaluation, matching attains an average 93.7% gain for all choice situations that involved 2 options. The strategy has an exponential decay of 2.6% per option, and achieves a 85.3% gain for 10-option decisions. In comparison, greedy subjects fall short when facing an appreciable number of options. This is because putting all effort into the seemingly most rewarding option leads to an opportunity loss (e.g., Figure 3B). The greedy strategy shows an exponential decay of 3.9% per option, and attains a mere 52.9% average gain in 10-option situations.
Matching is found to be particularly efficient in situations with diminishing returns, but provides high performance also in situations in which the law of diminishing marginal utility does not hold. To quantify how diminishing-increasing returns affect matching performance, the strength of diminishing returns is parameterized using the power function, , where is the tested parameter. For , the gain diminishes strongly with invested effort . On the other hand, for , gain increases linearly with effort and never saturates. Figure 2D shows the dependence of the total decision gain on this parameter. Matching is highly efficient—approaching perfect gain following a single evaluation of the decision options—when the choice options obey the law of diminishing marginal utility. This reproduces a previous, analytical study [13]. The degenerate, greedy strategy is particularly inadequate in this case: A greedy individual puts all effort or resources into an option whose payoff eventually saturates, thus experiencing an opportunity loss. As expected, the performance of matching/greedy behavior is lower/higher in choice situations in which the law of diminishing marginal utility does not hold. Nonetheless, matching can outperform greedy behavior also in these cases, when applied repeatedly . In particular, following the fourth evaluation, matching attains a 91.2% gain, compared to a 88.8% gain of the greedy approach. The same principal observation is made for functions other than the power function (Figure 4B).
Efficiency in stochastic environments
Matching provides a high-level strategy to effort allocation. Therefore, the results presented thus far are general to any task or choice situation that provides a certain kind of reward as a function of a certain kind of effort. This includes deterministic and stochastic choice situations. As a specific example of performance in stochastic choice environments, matching was deployed in environments in which rewards were provided according to the frequently investigated variable interval schedules [8, 13, 14, 33–35, 38, 42–47]. Matching and maximizing decision-makers integrated the individual binary rewards over a defined period of time using the commonly used reinforcement learning [38, 39, 48], and allocated effort according to the estimated utility of each option (Materials and Methods). The rate of reward occurrence for each option was again randomized across 1000 distinct choice situations, each involving 2, 3, and up to 10 options.
Matching was found to be rapid and efficient also in this case (Figure 3A). Following a single evaluation, matching harvested an average of 86.7% of the total possible gain. In comparison, maximization required an average of 29 option evaluations to reach equivalent performance. The greedy strategy produced a mere 54.1% gain.
Figure 3B exemplifies how matching achieves its high efficiency. Following an initial assessment of the relative worth of the available options (left panel), matching (green diamonds) allocates effort in proportion to the relative worth, i.e., . The second panel shows that this step readily results in harvesting the highest reward from the richest option (i.e., Option 3), as desired, while also gaining substantial reward from the relatively less valuable options. Repeated applications of matching (third and fourth panels) lead to additional gain. Maximization (gray squares) is found to require a large number of option evaluations to reach the performance level of matching (Figure 3A). The degenerate, greedy strategy (orange circles) settles with the option that appear richest at that moment, and thus misses the globally richest option.
Discussion
This study finds that matching provides effective behavioral allocations across a broad spectrum of decision situations, yielding high or near-optimal gain. Furthermore, the strategy is found to possess a notable strength—high efficiency. Matching approaches near-optimal gain following a single evaluation of the worth of the decision options (Figure 1, Figure 2, Figure 3). This finding holds across distinct choice situations (Figure 1), generalized forms of matching (Figure 2A), initial allocation of effort (Figure 2B), number of options (Figure 2C), the level of diminishing returns (Figure 2D), and stochastic environments (Figure 3).
The article finds that matching harvests on average around 90% of the maximum possible gain following a single evaluation of the worth of the decision options. Moreover, in economic settings, in which the worth of the options may be known or can be estimated (e.g., price), matching allows individuals to allocate effort (e.g., money) readily, without any prior sampling. Thus, matching is efficient, combining high performance with rapid evaluation.
The time to make a choice is a factor crucial to survival and rewarding behavior [49–53]. Indeed, most decisions involve a speed-performance or speed-accuracy tradeoff [54, 55]. This natural constraint limits the number of times individuals can assess and evaluate their options. Reward maximization, which provides the highest gain possible, is found to be taxing in this regard. For instance, Figure 1 and Figure 3 show that reward maximization, even when algorithmically efficient (Materials and Methods), would require a decision-maker to visit each option or food patch on average at least 16 times to make a decision of quality comparable to matching. In comparison, matching provides a single-step convergence on the optimum (Figure 1, Figure 2, Figure 3). Individuals using this efficient decision strategy may possess an evolutionary advantage.
Indeed, matching has been considered a candidate for an evolutionarily stable strategy [36, 37]. Referred to as the relative payoff sum in behavioral ecology, a population of individuals using this strategy is believed to be uninvadable by a mutant with a different strategy [36, 56]. A proof was attempted [36] but remains elusive [57]. The present study shows that this strategy is efficient, providing optimal or near-optimal gain after just one sampling of the worth of the choice options. Given that foraging efficiency constitutes an important optimization criterion in ecology [49, 50, 58], the result of this study supports the proposition that this strategy might be evolutionarily stable.
Matching is applicable both at the global level of the matching law and at the local level of decision strategies [38–40, 59]. For instance, matching was found to be used as a time-by-time decision strategy by non-human primates [38, 60]. The subjects allocated their relative effort—measured by the relative number of choices of an option—to the relative utility of each option. Moreover, the relative utility associated with the matching equation, termed fractional income, captured firing rates of neurons in the recorded brain region, the lateral intraparietal area. At the molecular level, matching is known to result from melioration [41], a strategy in which decision-makers gradually shift effort to the more valuable options. The repeated, iterative application of the matching rule shown here, i.e., (Figure 2B,D; Figure 3A,B), is an instance of melioration.
Humans and animals often exhibit under- or over-matching tendencies [25, 26]. The analyses performed here find that these slight deviations from the original, strict matching law [33] incur only a minor loss in performance (Figure 2A). Therefore, the principal discovery—the efficiency of matching as a strategy—also applies to the generalized form of matching. Generalized matching allows for the representation of value to be nonlinear, which is supported by several fields [21–23]. In this context, the article also shows that extreme forms of undermatching (i.e., indifference) and overmatching (i.e., greedy behavior) result in relatively low gain and thus can be considered maladaptive.
Decision-makers are known to weigh benefits against costs, or rewards against effort. Yet, how effort should be incorporated in current models of decision-making in economics [61], ecology [62], and psychology [63, 64] has been unclear. This article shows that individuals can make efficient cost-benefit decisions by adopting matching, i.e., evaluating the relative reward per relative effort of each option (in matching, for all options). This value formulation increases with the benefit of an option (), decreases with the cost of an option (), and the relativistic expression does not require absolute units and can thus be represented by neurons.
This study focused on the principles that individuals use to allocate their effort to their choice options (blue box in Figure 3C). The input to these choice rules—how subjects represent the worth of their options—is a separate question that has been addressed in neuroscience [38, 39, 48, 65]. Notably, in matching, the relativistic nature of and enables these variables to be subjective. So long as the underlying representations are comparable by the brain (e.g., through firing rates), the relativistic strategy provides a rapid and efficient allocation of the decision-maker’s resources.
In matching, the relativistic representation of utility involves divisive normalization. Divisive normalization is a common operation performed by neural circuits [66]. The specific form of this operation may be crucial for explaining attraction, similarity, and compromise effects observed in multi-alternative, multi-attribute decision environments [67, 68]. For instance, it has been found that a transformation of utilities by specific monotonic functions prior to divisive normalization can explain these behavioral effects parsimoniously [68]. Monotonic transformations and divisive normalization are performed by several kinds of feedforward and feedback neural circuits [66, 69].
In summary, this article evaluated the performance of matching across a wide range of choice environments, and uncovered a notable strength of this strategy across the environments—high efficiency. These quantitative findings may explain why biological decision-makers adopt matching so commonly.
Materials and Methods
Choice behavior
In general, decision-makers allocate their total effort, , to their options based on the options’ relative utilities or rewards. The effort can represent time, money, a physical or mental expenditure; generally, any resource that an individual allocates to her options in a given choice situation.
Maximization
Maximizing decision-makers allocate effort across their choice options to maximize the total reward or utility across their options.
Matching
In strict matching, decision-makers allocate their relative effort associated with each option in proportion to the associated relative utility or reward, , such that (Figure 3C). Generalized matching also follows , but the relative utility of each option is parameterized as , where the exponent . This parametrization captures subjective representation of utility [21–23].
Evaluations
Choice situations
The evaluations sampled a broad spectrum of utility-effort contingencies, i.e., functions. The functions were of distinct kinds, used distinct magnitudes, and distinct level of diminishing marginal utility (Figure 4A). Specifically, power, logistic, hyperbolic, and logarithmic functions in the forms , and , respectively, were evaluated. The logarithmic functions used in their argument to avoid negative utility, and in the denominator to implement logarithm base values . These reward-effort contingencies are encountered frequently in natural and societal settings [13]. The principal findings of this article are relatively insensitive to the particular functions evaluated.
Choice options
The study simulated a large number of choice situations to provide broadly applicable results. The situations featured options. For each option within the set of , the specific function (power, logistic, hyperbolic, or logarithmic), its magnitude , and the saturation parameter were all selected randomly. The function kind was selected from the set of the four. was selected randomly from a uniform distribution over the range from 0 to 1. The diminishing marginal utility level was selected randomly from the set {0.1, 0.2,…1.0}. The randomization was performed 1000 times for each , yielding a total of 9000 choice situations. This number of evaluations provided tight standard errors of the mean effects (Figure 1, Figure 2).
Stochastic choice situations
The stochastic environments used the same number of repetitions and the same number of options. Rewards were provided according to the commonly used variable interval schedules [8, 13, 14, 33–35, 38, 42–47]. In a variable interval schedule, a binary reward is assigned to an option randomly from the Poisson process at a specific reward rate [13]. The reward rate was again assigned randomly to each option, within the range from 0 to 1 per unit of time. A reward assigned to an option is available until the decision-maker harvests it. Harvesting events (i.e., the decision-maker’s behavior) are also binary and assigned to an option randomly from the Poisson process, now at the rate of the specific rate of effort (range from 0 to 1 per unit of time, subject to ). To estimate the reward rate from the noisy binary information, decision-makers must integrate the history of past rewards. A common approach for doing so is reinforcement learning, in which the integrated binary rewards are discounted exponentially [38, 39, 48], i.e., using a function . The learning rate is related to the time period of integration. This study used , for . The factor was set such that the integral of the exponential over was equal to one, i.e., . Each option evaluation operated on subsequent time period ; i.e., each option evaluation encountered a distinct distribution of rewards and behaviors, generated using the respective rates.
Maximization
In each of the choice situations, the maximizing subjects aimed to maximize their gain within a constrained distribution of effort devoted to each option. Specifically, the decision-makers maximized subject to The solution was provided using constrained nonlinear multivariable maximization (function fmincon in matlab, using the default interior-point algorithm; the specific form of the algorithm had only minimal effect on the results). The algorithm converged onto a solution in a finite number of options evaluations in each decision case (see Figure 1 and Figure 3A for the number of evaluations necessary to reach specific performance).
Matching
Matching decision-makers also distributed their effort across the choice options subject to finite effort,. Their total gain, , is reported as a percentage of the maximum possible gain harvested by maximization. It was found that matching individuals can use a single or just a small number of evaluations of the decision options. Specifically, they first sampled the worth of their options by allocating initial effort to each option. Subsequently, they applied the matching strategy i.e., . Thus, the decision-makers simply allocated their effort in proportion to an initially established worth of their options. In generalized matching, the proportionality still holds; nonlinearities are modeled using the exponent in the representation of relative utilities (see the Matching paragraph above). In Figure 2B, Figure 2D, and Figure 3, this strategy was applied iteratively using melioration, i.e., . Moreover, the analysis of Figure 2B also tested whether the strategy is sensitive to the kind of the initial effort distribution-equal or entirely random.
Acknowledgements
I thank Drs. Alex Kacelnik, Alasdair Houston, Leo Sugrue, and Julian Brown for comments on the manuscript. This work was supported by the NIH grant RF1NS128569 and NSF grant 2325125. The author has no conflict of interest.
Footnotes
Code availability
The simulation code is available and documented at neuralgate.org/download/EDM.
References
- 1.Samuelson PA. Foundations of economic analysis. Science and Society. 1948;13(1). [Google Scholar]
- 2.Rachlin H, Battalio R, Kagel J, Green L. Maximization theory in behavioral psychology. Behavioral and Brain Sciences. 1981;4(3):371–388. [Google Scholar]
- 3.Friedman M. Essays in positive economics. University of Chicago Press; 1953. [Google Scholar]
- 4.Becker G. The Economic Approach to Human Behavior. University of Chicago Press; 1978. [Google Scholar]
- 5.Binmore K. Rational Choice Theory: Sufficient but not necessary. American Psychologist. 1991;46(7):797–799. [Google Scholar]
- 6.Herrnstein RJ, Loveland DH. Maximizing and matching on concurrent ratio schedules 1. Journal of the experimental analysis of behavior. 1975;24(1):107–116. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7.Kagel JH, Battalio RC, Green L. Economic choice theory: An experimental analysis of animal behavior. Cambridge University Press; 1995. [Google Scholar]
- 8.Rachlin H. Economics of the matching law. Quantitative analyses of behavior. 1982;2:347–374. [Google Scholar]
- 9.Herrnstein RJ. The matching law: Papers in psychology and economics. Harvard University Press; 1997. [Google Scholar]
- 10.Poling A, Edwards TL, Weeden M, Foster TM. The matching law. The Psychological Record. 2011;61:313–322. [Google Scholar]
- 11.Wells VK, Foxall GR. Matching, demand, maximization, and consumer choice. The Psychological Record. 2013;63(2):239–258. [Google Scholar]
- 12.Davison M, McCarthy D. The matching law: A research review. Routledge; 2016. [Google Scholar]
- 13.Kubanek J. Optimal decision making and matching are tied through diminishing returns. Proceedings of the National Academy of Sciences. 2017;114(32):8499–8504. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Mazur JE. Optimization theory fails to predict performance of pigeons in a two-response situation. Science. 1981;214(4522):823–825. [DOI] [PubMed] [Google Scholar]
- 15.Houston AI, Trimmer PC, McNamara JM. Matching behaviours and rewards. Trends in Cognitive Sciences. 2021;25(5):403–415. [DOI] [PubMed] [Google Scholar]
- 16.Baum WM. On two types of deviation from the matching law: bias and undermatching. Journal of the experimental analysis of behavior. 1974;22(1):231–242. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Allen CM. On the exponent in the “generalized” matching equation. Journal of the experimental analysis of behavior. 1981;35(1):125–127. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18.MacDonall JS. Concurrent variable-ratio schedules: Implications for the generalized matching law. Journal of the Experimental Analysis of Behavior. 1988;50(1):55–64. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19.Horne PJ, Lowe CF. Determinants of human performance on concurrent schedules. Journal of the Experimental Analysis of Behavior. 1993;59(1):29–60. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20.Koffarnus MN, Woods JH. Quantification of drug choice with the generalized matching law in rhesus monkeys. Journal of the experimental analysis of behavior. 2008;89(2):209–224. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Kahneman D, Tversky A. Choices, values, and frames. American psychologist. 1984;39(4):341. [Google Scholar]
- 22.Fishburn PC. Multiattribute nonlinear utility theory. Management Science. 1984;30(11):1301–1310. [Google Scholar]
- 23.Glimcher PW, Dorris MC, Bayer HM. Physiological utility theory and the neuroeconomics of choice. Games and economic behavior. 2005;52(2):213–256. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24.Simon HA. Rational choice and the structure of the environment. Psychological review. 1956;63(2):129. [DOI] [PubMed] [Google Scholar]
- 25.William BM. Matching, undermatching, and overmatching in studies of choice. Journal of the Experimental analysis of Behavior. 1979;32(2):269–281. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Wearden J. Undermatching and overmatching as deviations from the matching law. Journal of the experimental analysis of behavior. 1983;40(3):332–340. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27.McNamara J, Houston A. The application of statistical decision theory to animal behaviour. Journal of Theoretical Biology. 1980;85(4):673–690. [DOI] [PubMed] [Google Scholar]
- 28.Hutchinson JM, Gigerenzer G. Simple heuristics and rules of thumb: Where psychologists and behavioural biologists might meet. Behavioural processes. 2005;69(2):97–124. [DOI] [PubMed] [Google Scholar]
- 29.Stephens DW, Krebs JR. Foraging theory. Princeton university press; 2019. [Google Scholar]
- 30.Herrnstein RJ. Derivatives of matching. Psychological Review. 1979;86(5):486–495. [Google Scholar]
- 31.Redmon WK, Lockwood K. The matching law and organizational behavior. Journal of Organizational Behavior Management. 1987;8(1):57–72. [Google Scholar]
- 32.Herrnstein RJ. Rational choice theory: Necessary but not sufficient. American Psychologist. 1990;45(3):356. [Google Scholar]
- 33.Herrnstein RJ. Relative and absolute strength of response as a function of frequency of reinforcement. Journal of the experimental analysis of behavior. 1961;4:267–72. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 34.Staddon J. Reinforcement as input: Cyclic variable-interval schedule. Science. 1964;145(3630):410–412. [DOI] [PubMed] [Google Scholar]
- 35.Herrnstein RJ. On the law of effect. Journal of the experimental analysis of behavior. 1970;13(2):243–266. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 36.Harley CB. Learning the evolutionarily stable strategy. Journal of theoretical biology. 1981;89(4):611–633. [DOI] [PubMed] [Google Scholar]
- 37.Tracy ND, Seaman JW Jr. Properties of evolutionarily stable learning rules. Journal of theoretical biology. 1995;177(2):193–198. [DOI] [PubMed] [Google Scholar]
- 38.Sugrue LP, Corrado GS, Newsome WT. Matching behavior and the representation of value in the parietal cortex. Science. 2004;304(5678):1782–7. [DOI] [PubMed] [Google Scholar]
- 39.Corrado GS, Sugrue LP, Seung HS, Newsome WT. Linear-Nonlinear-Poisson models of primate choice dynamics. Journal of the experimental analysis of behavior. 2005;84(3):581–617. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 40.Soltani A, Wang XJ. A biophysically based neural model of matching law behavior: melioration by stochastic synapses. J Neurosci. 2006;26(14):3731–44. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 41.Melioration Vaughan W., matching, and maximization. Journal of the Experimental Analysis of Behavior. 1981;36(2):141–149. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 42.Baum WM. Choice in free-ranging wild pigeons. Science. 1974;185(4145):78–79. [DOI] [PubMed] [Google Scholar]
- 43.Rachlin H. A molar theory of reinforcement schedules. Journal of the experimental analysis of behavior. 1978;30(3):345–360. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 44.Staddon JER. Limits to action: The allocation of individual behavior. Academic Press; 1980. [Google Scholar]
- 45.Davison M, McCarthy D. The matching law: A research review. Lawrence Erlbaum Associates, Inc; 1988. [Google Scholar]
- 46.Herrnstein RJ. The matching law: Papers in psychology and economics. Harvard University Press; 2000. [Google Scholar]
- 47.McDowell J. On the theoretical and empirical status of the matching law and matching theory. Psychological Bulletin. 2013;139(5):1000. [DOI] [PubMed] [Google Scholar]
- 48.Sutton RS, Barto AG. Introduction to reinforcement learning. MIT Press; 1998. [Google Scholar]
- 49.Railsback SF, Harvey BC. Analysis of habitat-selection rules using an individual-based model. Ecology. 2002;83(7):1817–1830. [Google Scholar]
- 50.Stephens DW. Decision ecology: foraging and the ecology of animal decision making. Cognitive, Affective, & Behavioral Neuroscience. 2008;8(4):475–484. [DOI] [PubMed] [Google Scholar]
- 51.Forstmann BU, Dutilh G, Brown S, Neumann J, Von Cramon DY, Ridderinkhof KR, et al. Striatum and pre-SMA facilitate decision-making under time pressure. Proceedings of the National Academy of Sciences. 2008;105(45):17538–17542. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 52.Gowaty PA, Hubbell SP. Reproductive decisions under ecological constraints: it’s about time. Proceedings of the National Academy of Sciences. 2009;106(supplement 1):10017–10024. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 53.Ahituv N, Igbaria M, Sella AV. The effects of time pressure and completeness of information on decision making. Journal of management information systems. 1998;15(2):153–172. [Google Scholar]
- 54.Chittka L, Skorupski P, Raine NE. Speed–accuracy tradeoffs in animal decision making. Trends in ecology & evolution. 2009;24(7):400–407. [DOI] [PubMed] [Google Scholar]
- 55.Heitz RP. The speed-accuracy tradeoff: history, physiology, methodology, and behavior. Frontiers in neuroscience. 2014;8:150. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 56.Hamblin S, Giraldeau LA. Finding the evolutionarily stable learning rule for frequency-dependent foraging. Animal Behaviour. 2009;78(6):1343–1350. [Google Scholar]
- 57.Houston AI, Sumida BH. Learning rules, matching and frequency dependence. Journal of Theoretical Biology. 1987;126(3):289–308. [Google Scholar]
- 58.Charnov EL. Optimal foraging, the marginal value theorem. Theoretical population biology. 1976;9(2):129–136. [DOI] [PubMed] [Google Scholar]
- 59.Loewenstein Y, Seung HS. Operant matching is a generic outcome of synaptic plasticity based on the covariance between reward and neural activity. Proceedings of the National Academy of Sciences. 2006;103(41):15224–15229. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 60.Sugrue LP, Corrado GS, Newsome WT. Choosing the greater of two goods: neural currencies for valuation and decision making. Nat Rev Neurosci. 2005;6(5):363–75. [DOI] [PubMed] [Google Scholar]
- 61.Buchanan JM. Cost and choice: An inquiry in economic theory. University of Chicago Press; 1978. [Google Scholar]
- 62.Zipf GK. Human behavior and the principle of least effort: an introduction to human ecology. Addison-Wesley Press; 1949. [Google Scholar]
- 63.Bial aszek W, Ostaszewski P, Green L, Myerson J. On four types of devaluation of outcomes due to their costs: Delay, probability, effort, and social discounting. The Psychological Record. 2019;69(3):415–424. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 64.Botvinick MM, Huffstetler S, McGuire JT. Effort discounting in human nucleus accumbens . Cognitive, affective, & behavioral neuroscience. 2009;9(1):16–27. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 65.Iigaya K, Ahmadian Y, Sugrue LP, Corrado GS, Loewenstein Y, Newsome WT, et al. Deviation from the matching law reflects an optimal strategy involving learning over multiple timescales. Nature communications. 2019;10(1):1–14. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 66.Carandini M, Heeger DJ. Normalization as a canonical neural computation. Nature Reviews Neuroscience. 2012;13(1):51–62. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 67.Noguchi T, Stewart N. In the attraction, compromise, and similarity effects, alternatives are repeatedly compared in pairs on single dimensions. Cognition. 2014;132(1):44–56. [DOI] [PubMed] [Google Scholar]
- 68.Dumbalska T, Li V, Tsetsos K, Summerfield C. A map of decoy influence in human multialternative choice. Proceedings of the National Academy of Sciences. 2020;117(40):25169–25178. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 69.Lek S, Delacoste M, Baran P, Dimopoulos I, Lauga J, Aulagnier S. Application of neural networks to modelling nonlinear relationships in ecology. Ecological modelling. 1996;90(1):39–52. [Google Scholar]