Skip to main content
Elsevier Sponsored Documents logoLink to Elsevier Sponsored Documents
. 2020 Jul;200:104244. doi: 10.1016/j.cognition.2020.104244

How decisions and the desire for coherency shape subjective preferences over time

Adam N Hornsby a,b,, Bradley C Love b,c
PMCID: PMC7315129  PMID: 32222615

Abstract

Recent findings suggest a bidirectional relationship between preferences and choices such that what is chosen can become preferred. Yet, it is still commonly held that preferences for individual items are maintained, such as caching a separate value estimate for each experienced option. Instead, we propose that all possible choice options and preferences are represented in a shared, continuous, multidimensional space that supports generalization. Decision making is cast as a learning process that seeks to align choices and preferences to maintain coherency. We formalized an error-driven learning model that updates preferences to align with past choices, which makes repeating those and related choices more likely in the future. The model correctly predicts that making a free choice increases preferences along related attributes. For example, after choosing a political candidate based on trivial information (e.g., they like cats), voters' views on abortion, immigration, and trade subsequently shifted to match their chosen candidate.

Keywords: Preference learning, Decision making, Intrinsic motivation, Free choice, Choice-induced preference change

1. Introduction

Every day, people are confronted with countless choices for which there is no objectively correct answer. These tend to be either preference judgments or moral decisions (Nakao, Ohira, & Northoff, 2012). Rather than being guided by extrinsic feedback, people choose these options freely for themselves, using their subjective preferences. We therefore refer to these choices as free choices. But how do people acquire these subjective preferences in the first place? The aim of this research is to understand the basis of people's subjective preferences.

We might learn about our own preferences in the same way we learn about others'; by observing and then rationalizing behaviour (Bem, 1967, Bem, 1972; Cushman, 2019). This is because we tend to lack introspective access to the mechanisms driving our behaviour, meaning that we have to post-rationalize in order to make sense of it (Bem, 1967, Bem, 1972; Mandler, 1975; Miller & Buckhout, 1973; Nisbett & Wilson, 1977). In a dramatic demonstration of this, people have been tricked by mischievous experimenters into justifying choices that they did not actually make (Hall, Johansson, Tärning, Sikström, & Deutgen, 2010; Sauerland et al., 2016; Strandberg, Sivén, Hall, Johansson, & Pärnamets, 2018). For example, after choosing their favourite flavour of jam in a taste test, participants were tricked into then justifying a different choice by experimenters, who covertly switched them mid-way through the experiment (Hall et al., 2010). Thus, rather than accessing the reasons for their choices directly, people seem to retrospectively infer them using evidence of their historic choices, even when that evidence is not valid.

As well as facilitating the inference of preferences, past choices also shape them. This has been demonstrated in studies of free choice, which show that after freely choosing an option, people tend to increase their subjective preference for it (Akaishi, Umeda, Nagase, & Sakai, 2014; Alos-Ferrer & Shi, 2012; Ariely & Norton, 2008; Brehm, 1956; Chammat et al., 2017; Cockburn, Collins, & Frank, 2014; Koster, Duzel, & Dolan, 2015; Miyagi, Miyatani, & Nakao, 2017; Nakao et al., 2016; Riefer, Prior, Blair, Pavey, & Love, 2017; Schonberg et al., 2014; Sharot, De Martino, & Dolan, 2009; Vinckier et al., 2019; Voigt, Murawski, & Bode, 2017). In the original free-choice paradigm, Brehm (1956) asked participants to rate a set of items (e.g. snack products), choose between two similarly rated options and finally to rate the full set again. Results showed that after making the forced choice, they had an increased preference for the chosen item on the final rating and a decreased preference for the rejected item. This is surprising, because it suggests that merely choosing or rejecting an option causes a person to update their subjective preference for it.

Although there has been some debate as to the validity of the free-choice paradigm in its original format, more recent studies have suggested that choice-based learning is real. For example, one particular concern about the original paradigm was that the first rating phase was noisy and therefore an imperfect measure of people's true preferences (Chen & Risen, 2010; Izuma & Murayama, 2013). However, researchers have since overcome this concern using various methodological adaptations, demonstrating choice-induced preference change does occur (Akaishi et al., 2014; Alos-Ferrer & Shi, 2012; Koster et al., 2015; Miyagi et al., 2017; Nakao et al., 2016; Schonberg et al., 2014; Sharot, Velasquez, & Dolan, 2010; Vinckier et al., 2019) and can be long-lasting (Sharot et al., 2009). For example, Sharot et al. (2010) asked participants to blindly choose between masked holiday destinations, which were only revealed to participants after one of two keys had been pressed. Subsequent ratings of those destinations were consistent with choice-induced preference change, even though choices had been randomly assigned to participants. More recent analyses of data collected from supermarket shoppers in-the-wild gives further credence to the claim that choices are self-reinforcing. In particular, a recent study of 283,000 British consumers found that their tendency to repeat a choice (i.e., exploit) strengthened as a function of the number of previous repetitions (Riefer et al., 2017). The consensus from studies inside and outside the laboratory is that free choices appear to be self-reinforcing, such that people come to prefer the options they choose.

Studies of free choice imply that people refer to past acceptances and rejections to infer what they like and dislike (Akaishi et al., 2014; Chammat et al., 2017; Cockburn et al., 2014; Izuma et al., 2010; Miyagi et al., 2017; Riefer et al., 2017). Yet, we know that people can also infer the value of things they've never tried. For example, one could infer that they would not enjoy sky diving, despite having never tried it. How do people do this? Rather than caching the value of individual options (e.g. beer varieties), people likely represent options and preferences within a shared, continuous, multidimensional space (e.g., varieties of hop, brand and brewing style) (Hornsby, Evans, Riefer, Prior, & Love, 2019). As depicted in Fig. 1, representing options and preferences in this way is beneficial, in that it provides a lower-dimensional learning problem and allows one to infer the relative value of any option in their environment, irrespective of whether it has been tried.

Fig. 1.

Fig. 1

Many popular models of decision-making cannot easily explain how people form strong subjective preferences for free choices made in the wild. a, In standard decision theory, it is assumed that preferences remain stable over time (Glimcher, 2009; von Neumann, Morgenstern, & Rubinstein, 1944). Indeed, for many researchers, the challenge is often learning what people's preferences are (e.g., by asking them to choose between options), rather than understanding how they became. b, Reinforcement Learning (RL) models contrast in that they assume preferences change over time. Specifically, RL agents learn to prefer actions with a higher expected reward, which they learn as they monitor extrinsic feedback from their environment (Sutton & Barto, 1998). While RL has been shown to account for many aspects of human learning well (for a review, see Daw & Tobler, 2013), these investigations have been largely confined to objective tasks, where there is a clear extrinsic signal steering the decision-maker. c, Studies of free choice — where there is no objectively correct answer — have shown that merely choosing an item increases one's preference for it. This has often been taken to imply that free-choices are self-reinforcing, so as to increase preferences towards the chosen option. Yet, caching values in this way would arguably not scale well, as it would require people to keep track of every item they'd ever tried. d, In this paper, we propose a novel theory of subjective preference formation and decision making that arguably scales better to free choices made in the real world. According to this theory, people encode preferences over attributes of free choices, such as the hop content in beer. After making a decision, they then update their preferences in the direction of the attributes that defined their prior choice, thereby increasing their preference for it, as well as for other options that have similar attributes.

Making free choices may therefore serve a considerably broader function than first thought, helping us to learn more deeply about ourselves and the world around us. Specifically, if options are represented within the same attribute space, then free choices may help to determine where one's preferences lie within that space. In this paper, we propose that the position of one's preferences is determined by a general, error-driven learning process, where the error term seeks to make the last choice more likely to repeat. As well as increasing the likelihood of past choices being repeated — as has been shown in real supermarket consumers (Riefer et al., 2017) — one should also increase their preferences for other options to the extent that they are similar to those previously chosen. While surprising, striving for internal coherency in this way may make sense in a world where choices can be evaluated across a multitude of different criteria.

We begin by demonstrating how the intrinsic desire to maximize coherency between past choices and present preferences can elicit strong subjective preferences in the absence of extrinsic reinforcement. In accordance with our proposed theory, we develop a computational cognitive model that learns preferences over choice attributes and uses past choices as the basis for updating them. We call it the Coherency Driven Choice (CDC) model. CDC is similar to models in the field of human category learning, which are primarily concerned with classification of items into a set of mutually exclusive categories via their attributes (e.g., using wings, beaks and feathers to describe birds) (Kruschke, 1992; Love, Medin, & Gureckis, 2004; Nosofsky, 2011). However, rather than updating based on corrective feedback, our model self-supervises using its past choices, thereby making them and similar options more likely to be sampled. Moreover, we use these mechanisms to make decisions that do not involve any fixed set of classes; the model chooses a set of items, which is not of fixed size. Through simulation, we show how this mechanism can drive complex, multidimensional preferences from free choices alone. Thus, error-driven, self-supervision helps the agent to maximize coherency between its preferences and choices over time. As a result, CDC can achieve a sense of order in environments where there are innumerable possible options and dimensions by which to score them.

After presenting this formal demonstration of our theory, we validate its predictions using a large-scale experiment of human participants. Chiefly, the error-driven nature by which CDC learns means that it will update its preferences in order to maximize the perceived contrast between accepted and rejected options. This is analogous to contrastive learning effects documented within the field of category learning, where the experience of contrasting category exemplars causes perceived category averages to drift apart and become idealized (Davis & Love, 2010). Results from Experiment 1 demonstrated that people update their preferences in a similar way following a choice. Specifically, participants were shown to prefer never-before-seen patterns if they happened to be on the back of a toy robot they had just designed. The more discriminating the pattern was to the initial robot, the more likely they were to prefer it.

Whereas Experiment 1 concerned preference formation in a novel and well-controlled domain, Experiment 2 evaluated the model's predictions in a domain in which people hold strong, preexisting preferences. In particular, Experiment 2 evaluated whether participants would retrospectively update their political beliefs following a vote. Results revealed that after choosing between two electoral candidates based on trivial grounds (e.g., whether they liked cats or dogs), participants were more likely to agree with a political belief later revealed by their chosen candidate, irrespective of whether that belief was traditionally left or right wing (e.g., pro-choice vs. pro-life abortion rights). Thus, these results support the key claims of our proposed theory, in that they suggest that people retrospectively update their preferences to be coherent with their past choices. Significantly, this even occurs in domains where people possess strong prior preferences that likely have strong subjective significance.

2. The coherency driven choice (CDC) model

We begin by formally describing our model of subjective preference learning and decision making.

Broadly speaking, the model works by maintaining an internal set of preferences and attention weights for attributes across choices. For example, all products in a supermarket can be described in terms of nutritional attributes such as salt, sugar and saturated fat content. Individuals will possess different preferences for those attributes, and pay differing levels of attention to them. These preferences and attention weights are used to determine how favourable a choice is at a given timepoint. In particular, the higher the attention-weighted similarity, the more likely it will be to choose that option. Our model can be thought of as an agent interacting with its environment. Much like a reinforcement learning agent, the model takes an action, observes its environment, updates its internal state and then repeats the process.

Here we introduce some important notation relevant to the model's decision making process. Note that vectors will now be denoted in bold lowercase letters and matrices in bold uppercase letters. We denote the observation of choices in the environment using the matrix O, which has a shape of N × M. Here, N denotes the number of choices available O = [o1, o2, …, oN]T at a given timestep. For simplicity of notation, we assume that the model must choose between two items at any one time (i.e. N=2). However — in principle — the model is not constrained to this. M denotes the number of attributes for each option. Thus, each column oi (i ∈ {1,...N}) is a vector of M attributes oi = [oi1, oi2, …, oiM]. Therefore, the element oij corresponds to the jth attribute (j ∈ {1,...M}) of the ith item.

2.1.1. Preference similarity

In order to determine the most appropriate choice, the model first calculates a probability over the available options observed in O using the preference vector p = [p1, p2, …, pM]T and the attention weight vector w = [w1, w2, …, wM]T. Each element of the preference and attention weight vectors pj and wj maps to an attribute j in the attribute vector oij.

After sampling an option from the environment, the agent must update the preference and attention weight vectors. We now discuss the process of computing probabilities, selecting actions and updating vectors in more detail.

2.1.2. Choice probabilities

In order to determine the probability of an action, the model calculates an attention-weighted similarity between the preference vector and each of the N = 2 item vectors oi within the observation matrix O. We denote the attention-weighted similarity as a(oi)

aoiγj=1Mwjoijpj212 (1)

where γ is a scaling hyperparameter. Note that this weighted Euclidean similarity term is very similar to the one used in the ALCOVE model of human category learning (Kruschke, 1992).

In order to determine the probability of selecting an option i, the attention-weighted similarity a(oi) is then fed into a softmax function

foiIiOsoftmaxaOi=expaoik=1Nexpaok (2)

Thus, the preference a(oi) for an option is a function of three things:

  • 1.

    The similarity (i.e. Euclidean distance) between the preference vector p and the choice attribute vector oi — The more similar the attributes and corresponding preference values, the more the model prefers that option

  • 2.

    The attention weight vector w — A higher degree of attention towards a similarity leads to a greater impact on the overall preference

  • 3.

    The scaling hyperparameter γ — The higher the γ, the higher the probability for selecting the preferred option (when using softmax action selection)

Similar to more traditional models, preferences are represented as ideal-points within a multidimensional space (Greenhoff & Mac Fie, 1994). However, unlike many of those methods, the model can have varying levels of attention to those according to the attention weights and — crucially — describes how preferences update over time as a consequence of decision making.

2.1.3. Action selection

Choices can be selected using one of the many popular strategies used in RL, such as ε-greedy, softmax action selection (Sutton & Barto, 1998) or more sophisticated directed-exploration strategies (e.g., uncertainty minimization). In each case, higher probabilities for choices (i.e. stronger preferences) increase the likelihood of exploiting that known favourite, rather than exploring disfavoured options. When using softmax selection specifically, the λ parameter can be thought of as determining the “fussiness” of the agent's choices, such that higher λ equates to a higher likelihood of choosing the favourite. We denote the choice made by the agent as c.

2.1.4. Updating preference and attention-weight vectors

Following an action, the agent must then update its preference and attention weight vectors. As discussed in the main text, a battery of psychological research has shown that — in subjective choice domains where there is no explicit feedback — preferences tend to follow choices. We therefore update the preference and attention weight vectors so as to maximize the likelihood of the previous choice. This contrasts sharply with traditional preference models, which seldom specify how preferences may change over time (DeSarbo & Kim, 2012; Greenhoff & Mac Fie, 1994).

The exact learning procedure used to update the preference and attention weight vectors is gradient descent on the cross-entropy loss, similar to that used during backpropagation and in the neural network literature generally (Goodfellow, Bengio, & Courville, 2016; Hinton, McClelland, & Rumelhart, 1986). During the learning procedure, an action is determined probabilistically using the softmax choice rule. After the action, the cross-entropy loss is calculated between the preference probabilities output by the model f(oi) and the actual choice c that was made.

lfOci=1N1c=ilogfoi (3)

After making an action, the preference and attention weights are updated so as to minimize the cross-entropy error. Concretely, they are updated proportionally to the negative of the error gradient.

We therefore use the following calculation to find the partial derivative of the preference vector p with respect to the cross-entropy loss:

lfOcpjγ2wji=1N1{c=i}foi1aoioijpj (4)

And the following calculation to find the partial derivative of the attention weight vector w with respect to the cross-entropy loss:

lfOcwjγ22i=1N1{c=i}foi1aoioijpj2 (5)

We then use these partial derivatives to update the existing preference and attention weight vectors using gradient descent. Concretely, we define the following update rules for the vectors p and w, respectively:

ppηplfOcpj (6)
wwηwlfOcwj (7)

where ηp and ηw represent the learning rates for the preference and attention weight vectors, respectively. As is standard during gradient descent, these learning rates scale the updated vectors and thus determine the magnitude of the update at a given time step.

3. Simulating free choices demonstrates how strong preferences can be learned over time

To illustrate how one could learn strong subjective preferences by virtue of their choice trajectory, we simulated the CDC model. In the simulated environment, there were two choice types that did not vary on the first dimension but varied significantly on the second dimension. This is analogous to choosing between two beer brands that are similar in taste but contrast in the colour of branding.

3.1. Method

3.1.1. Simulation

Observations in the environment were randomly sampled from distributions of two clusters. Choice type a had a cluster centroid of (0.2,0.8) while choice type b had a cluster centroid of (0.2,0.2). The standard deviation of each cluster was determined apriori to be 0.05; thereby making the two choice types linearly separable. A total of 500 observations were simulated.

The agent was initialised with middling preferences and attention weights across the two attributes of (0.5,0.5). It was also set to have a learning rate of 0.01, ε=0.05 and λ=1. Choices were simulated for 10,000 timesteps. At each timestep, the agent was forced to choose between two randomly-selected options from each choice type using an ε-greedy action selection strategy.

Choosing these parameters over a large number of timesteps helped the learning process to be smooth and stable across the agent's lifespan. This is therefore more representative of what might happen over a timescale of several years. However, it should be noted that the same learning trajectory documented below could be found using a larger learning rate over a shorter number of timesteps, though many such situations would not involve complete movement towards a goal.

3.2. Results and discussion

A visualization of this model and the results of this simulation are shown in Fig. 2.

Fig. 2.

Fig. 2

To illustrate how this theory can materialize in strong subjective preferences, we formalize it in a computational cognitive model, known as Coherency Driven Choice (CDC). a, CDC possesses a preference and attention-weight vector. When evaluating options, the model evaluates the similarity between its own attention-weighted preference and the attributes of the available options. The closer an option's attributes are to the model's attention-weighted preferences, the more likely the model is to select it. b, Following a choice, CDC updates its preference vector to make the past choice most likely using gradient descent, thereby moving towards the prior choice in attribute space. c, CDC also adjusts its attention weights to make prior choice more likely, effectively warping the preference space so that it becomes more sensitive to the attended attribute. d, As the model grows to prefer the options it initially chose, it tends to repeat the same choice type more over time. e, The model was simulated for 10,000 timesteps in a simple environment in which there were two choice types. Due to happenstance in the initial choices, the model began to prefer choice type a, adjusting its preferences and attention weights in favour of the attributes that make it unique. The preference history is coloured by attention weight ratio, such that blacker colors indicate a greater deal of attention paid towards attribute 2.

After simulating 10,000 forced-choices, CDC eventually came to posses preferences resembling the first choice type (i.e., type a). After an initial sequence of random actions, CDC began to quickly develop a preference in retrospect of them, and thus select choices consistent with this new-found preference. This process was then self-reinforcing, further strengthening preferences and therefore the likelihood for choice type a over time. Thus, when consumers become less likely to explore new products the more they exploit the same, it may be because their preferences are being self-reinforced by their past choices, as shown in this simulation (Riefer et al., 2017).

Uniquely, as CDC chose more of choice type a, the preferences and attention of the model moved most in favour of the attributes of the choice that made it unique. This is a known consequence of discrimination learning, but uniquely demonstrated here within the context of subjective preference change (Davis & Love, 2010; Ramscar, Yarlett, Dye, Denny, & Thorpe, 2010; Rescorla & Wagner, 1972). Exaggerating preferences in this way helped the model to maximize the perceived contrast between the accepted and rejected choices. This increases the likelihood of the past choice type being sampled again and reduces the likelihood of the rejected option being selected. Continuing the example introduced above, this suggests that a person will have an over-exaggerated preference towards the unique branding of their preferred beer, helping to retrospectively justify their apparent preference.

Of course, when formalizing a cognitive theory, one must make some assumptions about the world. For example, in the case presented here, one could argue that people do not always have complete knowledge of the attributes describing each option at the time of decision. Indeed, one may need to taste a product to know how salty it is, or they may vote for a political candidate before learning of their stance on free-trade. In this case, our model would simply use a placeholder for their preference on that particular attribute. Upon revelation of the attribute value for their prior choice (e.g., discovering how salty a snack was), they would then move their preference towards the attribute point in question.

In reality, people are unlikely to develop preferences as exaggerated as the ones learned in this simulation. This is because choices in real life are often more innumerate, multidimensional and overlapping. The high degree of similarity between options in the real world would cause our agent to explore more, and thereby develop a less extreme set of preferences. The environment may also provide additional sources of noise during decision making that elicit exploration and thus movement of preferences in new directions. For example, a preferred product may be out-of-stock in a store, a person may develop a new allergy or travel to a new country. Rather than attempt to account for the multitude of ways in which preferences change as a function of choices in-the-wild, the aim of this simulation was to highlight a important consequence of the coherency maximizing mechanism proposed here; namely, that preferences are updated in order to discriminate the choice just made.

4. People prefer novel patterns associated with their prior choice

The new account proposed in this paper suggests that by learning preferences over attributes of choices, people can generalize their preferences to novel options that are associated with ones previously chosen. Moreover, similar to error-driven models developed in the field of category learning (Davis & Love, 2010), it predicts that people will update their preferences in the direction of the most discriminative elements of their choice, in order to maximize the likelihood it being repeated.

Experiment 1 aimed to validate these predictions. Here, participants were asked to design a robot (the trial flow is depicted in Fig. 3a). They were then introduced to a second robot, before both turned around revealing previously unseen, randomly assigned patterns on their backs. Finally, participants were asked to choose between three patterns; one that was unique to the back of the robot they had previously designed (i.e. chosen unique), another that was shared across the backs of the two robots and a final pattern that was unique to the back of the robot they had not designed (i.e. non-chosen). It was hypothesized that — consistent with a discriminative account of learning — participants would prefer those novel patterns to the extent that they were uniquely associated with the robot they had just designed.

Fig. 3.

Fig. 3

a, The trial sequence observed by participants. First, participants were introduced to a robot and asked to design aspects of it. They were then introduced to a new robot, which was designed differently. These robots then turned around, revealing random patterns on their backs. One pattern was unique to the robot participants' had previously designed (i.e., chosen-unique), one was shared across both robots (i.e., shared) and the other was unique to the robot they had previously eschewed (i.e., non-chosen). Participants were then asked to rank the patterns in order of preference. b, This chart depicts the proportion that each image type was chosen as a first, second and third preference (with standard error bars). These results are consistent with the theory presented here, which predicts that choice-based learning generalizes most strongly to the unique, most-discriminating feature of the original choice (i.e., chosen unique), followed by the shared feature.

4.1. Method

4.1.1. Participants

One thousand and three participants were recruited from Amazon Mechanical Turk (mturk). Mturk (www.mturk.com) is generally known for being an inexpensive source of reliable human data (Crump, McDonnell, & Gureckis, 2013, though see McDuffie (2019) for a discussion on the possible limitations). Participants were required to have completed >1000 tasks (or HITs) and have an acceptance rate >95%. Participants had to be based in the US or Canada. Data from 37 participants were removed due to having an average response time two standard deviations greater than the mean (> 17.64 s). The mean age of the participants was 37.0 (SD = 11.4) and 50.2% were male. Participants were paid 50 for participating, which is typical for mturk (Horton & Chilton, 2010). Overall, the experiment took about 10 min.1

4.1.2. Design

The experiment used a between-groups design with 10 trials. Participants either chose a pattern that belonged to the back of the robot they previously designed (i.e., chosen-unique), the one they did not design (i.e., non-chosen) or was shared across both. The dependent variable was therefore the sum of the preferences over each of the choice types across each of the 10 trials.

4.1.3. Apparatus & stimuli

The study was designed using JavaScript and was accessed in a web browser. The task was presented in a 700 × 700 pixel screen. During the design phase of the experiment, participants responded by choosing attributes from a drop-down box. When stating their design preferences in the final phase, participants were asked to click each robot in order of their preference (from highest to lowest).

Each robot was designed using the Support Vector Graphics (SVG) format. These robots had a front — which could be designed by participants — and a back, which contained randomly-assigned patterns.

During the design phase, participants could design three aspects of the front of their robot; the stomach texture, the visor border colour and the eye colour. In each trial, participants could choose between two randomly-selected options for each design aspect.

In the final phase of each trial, participants were shown multicoloured geometric patterns. In total, there were three patterns per trial, randomly selected from one of 50 possible triplets. One pattern appeared on one random half of the back of the robot they had previously designed. Another pattern appeared on one random half of the back of the robot they had not designed. The final design was shared such that it appeared on both remaining halves of the two turned robots.

4.1.4. Procedure

Participants were initially briefed about the experiment in order to get their informed consent and asked to supply their age and gender. They were told that the task would take about 10 min and would require them to design robots and make choices. On each trial, participants were shown a robot with a randomly selected name and asked to design it. The hope was that by designing the robot, they would become more motivated about their choice, increasing ecological validity. They could choose between one of two randomly selected options for each of the three design aspects. After completing the design, participants were then introduced to a new, frowning anti-robot. This anti-robot was designed using all the attributes that the participant had previously eschewed. In addition, participants were warned that the anti-robot did not like the participants' design. They were then asked to reassure their own robot by clicking on it. Henceforth, we refer to this anti-robot as the non-chosen option. This is because it was uniquely designed using elements that had been explicitly rejected during the previous phase of the trial. After clicking on their own robot, both robots then turned around revealing randomized patterns on their backs (described above). The two robots then moved to the back of the screen. Three patterns — either shared across both or unique to one of the robots at the back of the screen — then appeared at the front of the screen and participants were asked to choose their favourite in order of preference. Participants completed 10 such trials with patterns, robot names, and other trial details randomized for each trial, they were debriefed, thanked for their participation and paid immediately.

4.2. Results and discussion

A repository for all data described in this paper is available on OSF ([dataset] Hornsby & Love, 2019).

The proportion of times each image type was selected as a first, second and third preference is pictured in Fig. 3b. As hypothesized, participants most preferred the unique chosen pattern, followed by the shared pattern and finally, the non-chosen pattern (omnibus non-parametric Friedman test of differences among repeated-measures χ2=137.48,p<0.001). Specifically, summed preferences for the chosen-unique patterns (Median = 9, IQR = 4.0) were stronger than that for the shared items (Median = 9, IQR = 3.0) (Wilcoxon signed-rank Z = − 2.91, p<0.005,r=0.09) and the non-chosen items (Median = 11, IQR = 5.0) (Z = − 12.08, p<0.001,r=0.39). A final test also revealed that preferences for the shared items were stronger than that for non-chosen items (Z = − 11.70, p<0.001,r=0.38).2

These results supported our two key hypotheses and therefore the key claims of our theory. Firstly, participants exhibited an increased preference for the chosen-unique and shared patterns, demonstrating that they generalized their preference learned from the initial choice to the novel patterns by virtue of their association. Consistent with the behaviour of our model formalization, people appear to generalize their preferences to novel items that share attributes with choices just made.

Secondly, participants demonstrated an increased preference for the chosen-unique pattern over the shared pattern and a reduced preference for the pattern unique to the rejected option. This is consistent with discriminative accounts of learning, in that it suggests that people update their preference towards attributes that discriminate their prior choices. Studies of human categorization have shown that experiencing contrasting category exemplars causes their perceived difference to drift apart and become idealized. For example, because people are used to contrasting diet foods with high calorie foods, they are more likely to suggest celery as a prototypical diet food, even though it is extreme for its category (Davis & Love, 2010). Choosing freely between options appears to have similar effects during preference learning, in that preferences update to maximize the perceived contrast between the accepted and rejected options. This functions to maintain coherence between past choices and present preferences.

5. People adjust their existing political beliefs to be consistent with their prior vote

The experimental results presented so far have provided controlled, experimental support for our error-driven account of subjective preference formation and decision making. Outside the lab however, people usually have strong prior preferences for options, which likely interacts with their intrinsic tendency to coherency maximize.

The aim of Experiment 2 was therefore to explore the extent to which people retrospectively update their existing preferences following a free choice. Specifically, we evaluated whether people would modify their political beliefs after a vote. Participants from the U.S. were shown two political candidates and asked to vote for one, based on some trivial attributes (e.g., whether they liked cats or dogs).3 This experimental procedure is depicted in Fig. 4a. Following the vote, these chosen and non-chosen political candidates revealed randomly assigned, opposing controversial beliefs on a particular topic. These beliefs were either traditionally left or right-wing. For example, if randomly assigned to the abortion issue, the chosen candidate would either show “Abortion: Pro-choice” or “Abortion: Pro-life”. Participants were then asked to state the extent to which they agreed with their chosen candidate's newly revealed belief using a slider. Finally, they were asked to state their preferred party out of the Democrats or Republicans.

Fig. 4.

Fig. 4

a, Participants were initially asked to vote for a candidate based on trivial characteristics. Both candidates then revealed opposing beliefs on a more controversial belief. Participants were then asked to rate how much they agreed with their candidate's newly revealed opinion. b, If voting for the candidate affects the extent to which participants agree with the later-revealed political belief, then one would expect participants to show different levels of agreement between groups, depending on whether their candidate revealed a traditionally left-wing view or a right-wing view. c, The results of the experiment are shown in kernel density plots (with rugplots below to depict the individual data points). These reveal that Republican-identifying participants were particularly prone to adjusting their preference to be in accordance with that revealed by their chosen candidate. d, Plotting effect sizes for each per-topic comparison shows that Republican-identifying participants were particularly prone to adjusting their preferences.

Our primary hypothesis was that people would show higher levels of agreement for the right-wing view (e.g., pro-choice abortion rights) if their chosen candidate revealed support for that view, compared to the left-wing view. This would provide support for the theory presented here and suggest that, following a vote, people are prone to adjust their political beliefs to be coherent with their chosen candidate.

While expecting that people would feel inclined to agree with their chosen candidate's newly revealed opinion, we also secondarily hypothesized that this would vary depending on individual differences. Specifically, we expected that different individuals would be more or less prone to updating their preferences retrospectively, due to differences in their longstanding political beliefs. Due to its simplicity, we used people's preferred political party affiliation as an indirect measure of these beliefs and subsequently predicted that Republican-identifying participants would be more likely to update their beliefs to be coherent with their initial vote. This was for two main reasons. Firstly, due to the broad and unpredictable nature of being in power, voters in support most likely have to accept more compromises and adjust their beliefs on occasion in order to remain coherent. Secondly, results from self-reported surveys suggest that conservatives value coherency comparatively more when making political decisions (e.g. they have been shown to value loyalty to the in-group and to authority comparatively more Mendez, 2017; Haidt & Graham, 2007; Jost, Glaser, Kruglanski, & Sulloway, 2003). While this is a somewhat secondary prediction that does not directly flow from our core theory, we used Experiment 2 as an initial exploration of group differences across domains.

5.1. Method

5.1.1. Participants

One thousand participants were recruited using mturk. All participants were required to have had at least 1000 of their previous tasks accepted on mturk, have a 95% mturk task acceptance rate and be based in the United States. All participants were paid 50 for participating. Of the participants, 47 were removed due to a lack of apparent concentration or understanding of the task. Specifically, 45 participants were removed for having response times more than two standard deviations from the mean for the vote response (n=26, >38.68s) and the slider response (n=19, >55.38s). A further 2 participants were removed for clicking >50 times during the whole experiment. Of the final 953 participants, 50.26% were male, 49.53% were female and the remainder specified as ‘other’. The mean age was 37.44 (SD = 11.67). Of the participants, 64.53% said they affiliated more closely with Democrats, whereas the remainder said they affiliate more closely with Republicans. Data was collected in late July 2018.

5.1.2. Design

The experiment used a 2 × 2 between-groups design. The experiment involved one trial. Participants either chose a candidate that — previously unbeknownst to them — revealed a more left-wing view or a more right-wing view. Furthermore, participants either affiliated more closely with Democrats or Republicans. The dependent variable was the slider value of the participant, normalized by political direction. Specifically, these values ranged from 1 to 100, where higher values indicated more agreement with the right-wing view and lower values indicated more agreement with the more left-wing view. Note that the initial vertical location of the candidates, the allocation of neutral and controversial attributes, and the final horizontal locations of the candidates were all fully randomized.

5.1.3. Apparatus & stimuli

The study was designed using JavaScript and was accessed in a web browser. The task was presented in a 1000 pixel wide screen. Participants were shown two of a possible four faces taken from the Chicago Face Database (Ma, Correll, & Wittenbrink, 2015).4 To control for noise emerging from other, well-documented biases, all faces were controlled to be male, Caucasian and ranging between the ages of 37 and 43. Faces were randomly assigned one of four names. These were either James Smith, Michael Johnson, John Williams, Graham Brown. These names were sourced from a list of the most popular first and second Caucasian names in the United States.

The neutral opinions of participants are shown in Table 1. The controversial political opinions of participants are shown in Table 2.

Table 1.

A summary of the neutral statements shown to participants.

Topic Statements
Pets “I like cats”
“I like dogs”
Sport “I'm a baseball fan”
“I'm a basketball fan”
Food “My favourite Italian food is pizza”
“My favourite Italian food is spaghetti”
Table 2.

A summary of the controversial statements revealed by candidates following a vote.

Topic Left-wing Right-wing
Abortion “Abortion: Pro-choice” “Abortion: Pro-life”
Immigration “I support policies that would increase immigration” “I support policies that would decrease immigration”
Trade “I support tariffs on imports” “I support free trade”

5.1.4. Procedure

Participants were initially briefed about the purpose of the experiment in order to get their informed consent. They were told that the task would take about 5 min to complete and would require them to vote for political candidates. After the briefing, participants began the trial. To first ensure that the participants were alert, they were told to read a short experimental briefing. Within this briefing was the instruction to click on the name of the university (which was displayed at the top of the screen). After clicking this, a new panel was revealed displaying some text (“It's time to vote for a candidate! Please click on a candidate to vote”) and two political candidates in gray cards; one above the other. Within each card and underneath the candidate's photographs were their name and then a list of “My opinions”. These lists initially displayed opposing neutral opinions selected from the same neutral topic, as shown in Table 1. The assignments of photographs, names, neutral topics and vertical alignment were all randomized. Participants were asked to vote for a candidate using this neutral information by clicking on their card. After voting, the cards then moved to a left or right position on the screen, aligning horizontally with each other. This horizontal allocation was also randomized. The voted and non-voted candidates then immediately revealed opposing political opinions from the same topic in their list of opinions. A blue “New” button drew attention to this newly-revealed belief. The window then moved down to reveal a new section, containing some text and a slider. The text informed participants in bold lettering that “Your candidate has revealed a new opinion above! How much do you agree with your candidate's newly revealed opinion?” Participants were prompted to use the new slider to state the extent to which they agreed with the candidates' newly revealed opinions. Small avatars of the candidate's faces were shown on either side of the slider, with a gray tick or white cross below the ones that were previously chosen, respectively. There were also prompts below each avatar reminding participants of the newly revealed controversial opinion of the candidate. Note that the slider was not initialized with any starting value. Sliding closer towards the chosen candidates indicated higher levels of agreement with that candidate. The slider had 100 possible positions. The photographs, names, neutral and political opinions and left/right direction of movement were all randomly assigned between participants.

After confirming the slider position, participants were asked to state which party they most affiliated with (Democrat or Republican), their gender and their age. Following task completion, they were thanked for their participation and paid immediately.

5.2. Results and discussion

In support of the primary hypotheses, results revealed that participants' stated degree of belief was significantly influenced by the randomly-assigned belief revealed by their chosen candidate (non-parametric two-way analysis of variance (ANOVA) (F (1, 952) = 28.89, p < 0.001, CL = 0.563). In particular, those that voted for a candidate that later revealed a right-wing opinion (Median = 56.0, IQR = 76.00) agreed an average of 43.59% more with the right-wing view compared to if the candidate later revealed a left-wing view (Median = 39.0, IQR = 69.75). This suggests that choosing a political candidate based on initially trivial characteristics made participants more likely to agree with that candidate's later-revealed controversial opinion, irrespective of whether the person self-identified as a Democrat or Republican.

In support of the secondary hypothesis, there was also a significant interaction between party affiliation and the degree of agreement with the right-wing view (F(1, 952) = 11.77, p < 0.001). A first Mann-Whitney U test looking at the responses of self-identifying Democrats revealed that they only slightly increased their level of agreement for the right-wing view if their candidate revealed a right-wing view (Median = 40.0, IQR = 78.00) compared to if it revealed a left-wing view (Median = 31.5, IQR = 65.75); this difference was not significant (U = 43,584.5, p = 0.054), and had a very low effect size (CL = 0.518). In contrast, the second test looking at Republican-identifying participants revealed a larger effect, in that — if one's candidate later revealed a more right-wing opinion — they were 65.66% more in agreement with the the right-wing view (Median = 82.0, IQR = 59.75) compared to when the candidate revealed a more left-wing view (Median = 49.50, IQR = 70.25) (U = 10,131.5, p < 0.001, CL = 0.629).5 Thus, Republican-identifying participants appeared to be particularly prone to adjusting their beliefs so as to be consistent with their last choice, even when this choice was based on trivial grounds (e.g., the fact that their candidate liked cats).

It is notable that self-identified Republicans were particularly willing to adjust their preferences to be coherent with their prior choices. For example, their median support for pro-life abortion policies was 60% higher when their chosen candidate later expressed support for pro-life policies (Median = 96.0, IQR = 61.50) compared to pro-choice policies (Median = 60.0, IQR = 88.00) (the remaining within-topic comparisons are depicted in Fig. 4c and 4d, and described in the supplemental). One possibility is that only Republican-identifying participants adjust their preferences to be consistent with a choice, across all domains. To confirm that this was not the case, we replicated the first experiment, asking about party affiliation at the end of the study. Results revealed that both self-identifying Democrat and Republican participants exhibited preferences for the final pattern in ways consistent with the reported findings above (see the supplemental for details). Thus, a more likely explanation of our results is that there are group and domain differences guiding the extent to which people learn from their past choices. For example — as discussed in the introduction to this section — differences in people's belief systems may give rise to different propensities to adjust preferences following a choice. Clearly these topics are deserving of future investigation in light of the results presented here.

6. General discussion

In this article, we proposed a novel account of subjective preference formation and decision making. In our model, choice options and preferences are represented in a common continuous, multi-dimensional space. When people choose options within this space, their preferences are updated to increase the likelihood of their previous choices. The objective of decision making is therefore to reduce the error between one's past choices and present preferences; we refer to this general mechanism as coherency maximization. Consistent with patterns of repeat-purchasing observed in supermarket consumers (Riefer et al., 2017) and studies of choice-induced preference change (Akaishi et al., 2014; Alos-Ferrer & Shi, 2012; Ariely & Norton, 2008; Brehm, 1956; Chammat et al., 2017; Cockburn et al., 2014; Koster et al., 2015; Miyagi et al., 2017; Nakao et al., 2016; Schonberg et al., 2014; Sharot et al., 2009; Vinckier et al., 2019; Voigt et al., 2017), coherency maximization boosts the likelihood of past choices being repeated by shifting preferences towards the chosen item and away from rejected alternatives. As was shown in a simple simulation, this mechanism can drive strong subjective preferences in the absence of extrinsic feedback. Thus, our model is well-suited to navigating the complex choices encountered in everyday life.

Following from this account, one would expect preferences to be exaggerated towards attributes that favoured the past choice and diminished towards attributes associated with rejected choices. Results from the first experiment showed that preferences are updated in accordance with these predictions. Concretely, the more uniquely associated a novel pattern was to the back of a toy robot they had previously designed, the more likely participants were to prefer it. Given that we tend to lack direct introspective access to the mechanisms driving our behaviour, post-rationalizing choices may be a reasonable way to make sense of ourselves and the world around us (Bem, 1967, Bem, 1972; Cushman, 2019; Mandler, 1975; Miller & Buckhout, 1973; Nisbett & Wilson, 1977). Our model shows that simple learning processes can achieve similar ends by updating underlying preferences to align with a choice.

These learning rules can be seen as reflecting an internal drive for internal consistency, which is pivotal to rational models of decision making (Savage, 1954). For example, it is often assumed that subjective choices should be stochastically transitive (Schultz, Dayan, & Montague, 1997). When preferences are not transitive, people can become liable to manipulation (or “dutch booking”) (Davidson, McKinsey, & Suppes, 1955). For example, if out of three beers, a person prefers beer A over B, beer B over C but would rather have C than A, their preferences are cyclical. This person could be tricked into paying for a series of costly trades in which the drinker ended with his original beer. In subjective domains, internal consistency may be at times the only rational strategy that is feasible.

The second experiment explored how the tendency to maximize coherency interacted with people's prior preferences. The results indicated that after voting for a political candidate based on trivial criteria (e.g., the candidate likes cats), participants were more likely to agree with a controversial opinion later revealed by the candidate, such as their stance on abortion rights. This held up as a main effect, supporting our claim that people retrospectively update their preferences over attributes of their past choices to make them more likely. The fact that this was particularly pronounced for Republican-identifying participants supported our secondary hypothesis that this mechanism can vary between groups, consistent with previous studies demonstrating individual differences in choice-induced preference change (Cockburn et al., 2014). However, while individual differences may partly explain our results, it is likely that other variables outside the scope of our theory also influence decisions, such as how important affiliation is to different groups. Specifically, while Republican-identifying participants adjusted their preferences more than their Democrat-identifying counterparts in the second experiment, a replication of the first experiment (see Supplement) revealed no difference between the groups when choosing between novel options. In future research, the prominence of choice-induced preference change within certain groups or domains could be estimated by fitting the CDC model to different decision tasks.

If people update their preferences to be coherent with their past choices, then why might they also feel motivated to explore? A recent analysis of consumers' take-away purchases suggested that they were more likely to try a different restaurant after a positive experience rather than repeat it again (Schulz et al., 2019), suggesting that people may also explore to reduce uncertainty about their environment. In the simulation, we used a stochastic, undirected exploration policy (i.e. e-greedy). However, CDC could be adapted to use a more sophisticated, directed exploration strategy, such as uncertainty minimization. This is because CDC considers exploration and coherency maximization as theoretically and mechanistically independent. Understanding where and when people explore is an open question in the literature (for a review, see Hills et al., 2015), meaning that such adaptations would need to be evaluated with scrutiny. In the future, we hope to further understand how people trade-off this need to explore with the desire to coherency maximize.

In this article, we have assumed that preferences are adjusted following a choice. However, preference change has also been shown to occur “online” during a choice (Akaishi, Kolling, Brown, & Rushworth, 2016; Niv et al., 2015; Schonberg et al., 2014; Voigt, Murawski, Speer, & Bode, 2019). A recent study by (Voigt et al., 2019) found that choice-induced preference changes only occurred for choices that were remembered. Both online and post-choice induced preference change mechanisms could co-exist. Though this would highlight a future area of development for the CDC model. One such modification would be to adapt the role of the attention weights during a choice. For example, their influence could be magnified in cases where the model was more familiar with or remembered the attributes of the choice. The preference vectors and attention weights would then account for both online and post-decisional effects of preference change, respectively.

Although our studies involved brief decision-making episodes, the basic mechanisms considered here may also apply at longer timescales. Indeed, this work was partially motivated by Riefer et al.'s (Riefer et al., 2017) discovery of self-reinforcing purchasing patterns in supermarket consumers, which extended over several months. If CDC's predictions for how preferences change held over extended periods of time, the practical consequences and possibilities for behavioural change would be substantial.

For instance, CDC predicts that purchasing the same type of food should increase preference for associated attributes. This can be problematic in cases where the food is unhealthy (e.g., high in sodium or saturated fat). For example, studies of nutrition have shown that repeated exposure to a particular ingredient (e.g., sodium) increases one's desire and lowers their sensitivity to it, making it difficult for them to adjust when they go on a diet (Bertino, Beauchamp, & Engelman, 1982; Mattes, 1993; Liem & de Graaf, 2004). The advent of targeted recommender systems means that these cyclical effects may be being perpetuated further. While recommender systems often reduce diversity of alternatives in the environment (Pariser, 2011), coherency maximization causes preferences to become less diverse, thereby perpetuating the problem. Rather than blindly targeting people based on their previous consumption, marketers could incorporate external objectives to their targeting algorithms. For example, our theory predicts that recommending healthy alternatives that are similar to people's existing preferences could lead to long-term improvements in their choices.

The modeling approach presented here is readily extended to account for the richness of people's preferences. Unlike how CDC was formalized here, people are unlikely to have a single preference across each respective attribute. For example, while a foodie might prefer expensive, locally produced foods, they may also be happy to watch affordable, mass produced television. People may have different preferences for the same attribute (e.g., cost) depending on context. Fortunately, it would be straightforward to extend CDC to have multiple preference vectors to capture this context dependence, similar to how models of human category learning possess multiple clusters in which only the most contextually relevant one is updated during a learning episode (Love et al., 2004). Such a model would cast coherency maximization as a process that occurs within a domain (e.g., food, entertainment, etc.) as opposed to globally.

The model presented here may also be extended to account for some well-known decision biases. For example, CDC predicts that people will prefer novel options when they are similar to options that have repeatedly tried in the past. Thus, it may be able to account for the mere-exposure effect, where people have been shown to prefer options by virtue of their familiarity (Zajonc, 1968). Similarly — though only distantly related to the free choices described here — it is possible that CDC could be adapted to account for anchoring effects. For example, forcing CDC to update its preferences towards a given anchor would cause it to become more favourable towards related options (see attitudinal change accounts of anchoring for related arguments, e.g., Wegener, Petty, Detweiler-Bedell, & Jarvis, 2001). While such relationships are speculative at this stage, they indicate a general, coherency-driven learning mechanism may underpin several, well-known sequential choice biases.

In conclusion, preferences and choices can be characterized as existing within a common space. While we prefer options that match our preferences, we also appear to engage in error-driven learning to update our preferences to accommodate our past choices. Because preference and choice representation lie in a shared multidimensional space, the choices we make have consequences (i.e., spillover effects) for related future choices. For example, people may be prone to agree with a controversial opinion held by a political candidate, by virtue of the fact that they voted for them. Although this behaviour may appear irrational, being internally coherent may be the best we can hope for in complex, subjective domains. Being aware of these coherency maximizing dynamics may make it possible for people to ameliorate some of the potentially harmful consequences. For example, if a voter chooses a candidate based on tax policy, perhaps being mindful of that fact will make the voter less likely to reflexively adopt their candidate's positions on unrelated issues. Likewise, these same coherency maximizing principles could be incorporated into recommender systems to help consumers achieve some positive goal, such as eating more healthily.

CRediT authorship contribution statement

Adam N. Hornsby: Conceptualization, Methodology, Software, Formal analysis, Investigation, Writing - original draft, Visualization, Writing - review & editing, Resources, Validation. Bradley C. Love: Conceptualization, Writing - review & editing, Supervision, Project administration, Funding acquisition.

Acknowledgments

Acknowledgements

We would like to thank Sébastien Ohleyer, Mina Stojanovic, Gareth Brown, Olivia Guest, Brett Roads and Sebastian Bobadilla Suarez for their help and feedback. A. N. H. is supported by Dunnhumby and the Royal Commission for the Exhibition of 1851. B.C.L. is supported by the Royal Society Wolfson Fellowship 1830295, Wellcome Trust Senior Investigator Award WT106931MA, and National Institute of Child Health and Human Development Grant 1P01HD080679.

Declaration of competing interest

The authors declare no conflicts of interest.

Footnotes

1

Both experiments presented here were in compliance with UCL's code of ethics

2

All Wilcoxon-signed rank tests were evaluated against a Holm-Bonferroni corrected alpha value for multiple comparisons

3

More information about the experimental design and stimuli is available in the methods section

4

These faces can be identified in the Chicago Face Database using the target codes WM-023, WM-221, WM-223 and WM-248

5

For all multiple comparisons reported in this paper, significance values were compared against a Holm-Bonferroni corrected α value

Appendix A

Data will be available at https://osf.io/5bvmp/ after publication. Code is available at https://github.com/anhornsby/coherent-representations. A Docker container for replicating our simulations and statistical analyses will be available at https://hub.docker.com/r/adamnhornsby/coherent-representations after publication. Supplementary data to this article can be found online at https://doi.org/10.1016/j.cognition.2020.104244.

Appendix A. Supplementary data

Supplementary material

mmc1.pdf (177.9KB, pdf)

References

  1. Akaishi R., Kolling N., Brown J.W., Rushworth M. Neural mechanisms of credit assignment in a multicue environment. Journal of Neuroscience. 2016;36(4):1096–1112. doi: 10.1523/JNEUROSCI.3159-15.2016. [DOI] [PMC free article] [PubMed] [Google Scholar]
  2. Akaishi R., Umeda K., Nagase A., Sakai K. Autonomous mechanism of internal choice estimate underlies decision inertia. Neuron. 2014;81(1):195–206. doi: 10.1016/j.neuron.2013.10.018. [DOI] [PubMed] [Google Scholar]
  3. Alos-Ferrer C., Shi F. Choice-induced preference change and the free-choice paradigm: A clarification. SSRN. 2012 doi: 10.2139/ssrn.2062507. [DOI] [Google Scholar]
  4. Ariely D., Norton M.I. How actions create – Not just reveal – Preferences. Trends in Cognitive Sciences. 2008;12(1):13–16. doi: 10.1016/j.tics.2007.10.008. [DOI] [PubMed] [Google Scholar]
  5. Bem D.J. Self-perception: An alternative interpretation of cognitive dissonance phenomena. Psychological Review. 1967;74(3):183–200. doi: 10.1037/h0024835. [DOI] [PubMed] [Google Scholar]
  6. Bem D.J. In: Self-perception theory. Berkowitz L., editor. Vol. 6. Academic Press; 1972. pp. 1–62. [DOI] [Google Scholar]
  7. Bertino M., Beauchamp G.K., Engelman K. Long-term reduction in dietary sodium alters the taste of salt. The American Journal of Clinical Nutrition. 1982;36(6):1134–1144. doi: 10.1093/ajcn/36.6.1134. [DOI] [PubMed] [Google Scholar]
  8. Brehm J.W. Postdecision changes in the desirability of alternatives. Journal of Abnormal and Social Psychology. 1956 doi: 10.1037/h0041006. [DOI] [PubMed] [Google Scholar]
  9. Chammat M., Karoui I.E., Allali S., Hagège J., Lehongre K., Hasboun D.…Naccache L. Cognitive dissonance resolution depends on episodic memory. Scientific Reports. 2017;7(41320) doi: 10.1038/srep41320. [DOI] [PMC free article] [PubMed] [Google Scholar]
  10. Chen M.K., Risen J.L. How choice affects and reflects preferences: Revisiting the free-choice paradigm. Journal of Personality and Social Psychology. 2010;99:573–594. doi: 10.1037/a0020217. [DOI] [PubMed] [Google Scholar]
  11. Cockburn J., Collins A.G.E., Frank M.J. A reinforcement learning mechanism responsible for the valuation of free choice. Neuron. 2014;83(3):551–557. doi: 10.1016/j.neuron.2014.06.035. [DOI] [PMC free article] [PubMed] [Google Scholar]
  12. Crump M.J.C., McDonnell J.V., Gureckis T.M. Evaluating amazon’s Mechanical turk as a tool for experimental behavioral research. PLoS One. 2013;8(3):1–18. doi: 10.1371/journal.pone.0057410. [DOI] [PMC free article] [PubMed] [Google Scholar]
  13. Cushman F. Rationalization is rational. Behavioral and Brain Sciences. 2019:1–69. doi: 10.1017/S0140525X19001730. [DOI] [PubMed] [Google Scholar]
  14. Davidson D., McKinsey J.C.C., Suppes P. Outlines of a formal theory of value, i. Philosophy of Science. 1955;22(2):140–160. [Google Scholar]
  15. Davis T., Love B.C. Memory for category information is idealized through contrast with competing options. Psychological Science. 2010;21(2):234–242. doi: 10.1177/0956797609357712. [DOI] [PubMed] [Google Scholar]
  16. Daw N., Tobler P. Elsevier; 2013. Value learning through reinforcement: The basics of dopamine and reinforcement learning. [DOI] [Google Scholar]
  17. DeSarbo W.S., Kim S. Quantitative modelling in marketing and management. 2012. A review of the major multidimensional scaling models for the analysis of preference/dominance data in marketing; pp. 3–25. [DOI] [Google Scholar]
  18. Glimcher P.W. Chapter 32- choice: Towards a standard back-pocket model. In: Glimcher P.W., Camerer C.F., Fehr E., Poldrack R.A., editors. Neuroeconomics. Academic Press; London: 2009. pp. 503–521. [DOI] [Google Scholar]
  19. Goodfellow I., Bengio Y., Courville A. MIT Press; 2016. Deep learning. [Google Scholar]
  20. Greenhoff K., Mac Fie H.J.H. Measurement of food preferences. Springer US; Boston, MA: 1994. Preference mapping in practice; pp. 137–166. [DOI] [Google Scholar]
  21. Haidt J., Graham J. When morality opposes justice: Conservatives have moral intuitions that liberals may not recognize. Social Justice Research. 2007, Mar 01;20(1):98–116. doi: 10.1007/s11211-007-0034-z. [DOI] [Google Scholar]
  22. Hall L., Johansson P., Tärning B., Sikström S., Deutgen T. Magic at the marketplace: Choice blindness for the taste of jam and the smell of tea. Cognition. 2010;117(1):54–61. doi: 10.1016/j.cognition.2010.06.010. [DOI] [PubMed] [Google Scholar]
  23. Hills T.T., Todd P.M., Lazer D., Redish A.D., Couzin I.D., Group, C.S.R Exploration versus exploitation in space, mind, and society. Trends in Cognitive Sciences. 2015;19(1):46–54. doi: 10.1016/j.tics.2014.10.004. [DOI] [PMC free article] [PubMed] [Google Scholar]
  24. Hinton, G. E., McClelland, J. L., & Rumelhart, D. E. (1986). Parallel distributed processing: Explorations in the microstructure of cognition, vol. 1 In D. E. Rumelhart, J. L. McClelland, & C. PDP Research Group (Eds.) (pp. 77–109). Cambridge, MA, USA: MIT Press.
  25. Hornsby A., N., Evans T., Riefer P., S., Prior R., Love B., C. Conceptual Organization is Revealed by Consumer Activity Patterns. Computational Brain and Behavior. 2019 doi: 10.1007/s42113-019-00064-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  26. Hornsby A.N., Love B.C. How decisions and the desire for coherency shape subjective preferences over time. 2019. https://osf.io/5bvmp/ [DOI] [PMC free article] [PubMed]
  27. Horton J., Chilton L. 2010. The Labor Economics of Paid Crowdsourcing. CoRR, abs/1001.0627. [DOI] [Google Scholar]
  28. Izuma K., Matsumoto M., Murayama K., Samejima K., Sadato N., Matsumoto K. Neural correlates of cognitive dissonance and choice-induced preference change. Proceedings of the National Academy of Sciences. 2010;107(51):22014–22019. doi: 10.1073/pnas.1011879108. [DOI] [PMC free article] [PubMed] [Google Scholar]
  29. Izuma K., Murayama K. Choice-induced preference change in the free-choice paradigm: A critical methodological review. Frontiers in Psychology. 2013;4:41. doi: 10.3389/fpsyg.2013.00041. [DOI] [PMC free article] [PubMed] [Google Scholar]
  30. Jost J., Glaser J., Kruglanski A., Sulloway F. Political conservatism as motivated social cognition. Psychological Bulletin. 2003;129(3):339–375. doi: 10.1037/0033-2909.129.3.339. [DOI] [PubMed] [Google Scholar]
  31. Koster R., Duzel E., Dolan R.J. Action and valence modulate choice and choice induced preference change. PLoS One. 2015;10(3):1–10. doi: 10.1371/journal.pone.0119682. [DOI] [PMC free article] [PubMed] [Google Scholar]
  32. Kruschke J.K. ALCOVE: An exemplar-based connectionist model of category learning. Psychological Review. 1992;99:22–44. doi: 10.1037/0033-295X.99.1.22. [DOI] [PubMed] [Google Scholar]
  33. Liem D.G., de Graaf C. Sweet and sour preferences in young children and adults: Role of repeated exposure. Physiology Behavior. 2004;83(3):421–429. doi: 10.1016/j.physbeh.2004.08.028. [DOI] [PubMed] [Google Scholar]
  34. Love B.C., Medin D.L., Gureckis T.M. SUSTAIN: A network model of category learning. Psychological Review. 2004;111(2):309–332. doi: 10.1037/0033-295X.111.2.309. [DOI] [PubMed] [Google Scholar]
  35. Ma D.S., Correll J., Wittenbrink B. The Chicago face database: A free stimulus set of faces and norming data. Behavior Research Methods. 2015;47(4):1122–1135. doi: 10.3758/s13428-014-0532-5. [DOI] [PubMed] [Google Scholar]
  36. Mandler G. Consciousness: Respectable, useful, and probably necessary. Mattes, R. D. (1993, 03). Fat preference and adherence to a reduced-fat diet. The American Journal of Clinical Nutrition. 1975;57(3):373–381. doi: 10.1093/ajcn/57.3.373. [DOI] [PubMed] [Google Scholar]
  37. Mattes R.D. Fat preference and adherence to a reduced-fat diet. The American Journal of Clinical Nutrition. 1993;57(3):373–381. doi: 10.1093/ajcn/57.3.373. [DOI] [PubMed] [Google Scholar]
  38. McDuffie D. Using amazon’s mechanical turk: Benefits, drawbacks, and suggestions. APS Observer. 2019;32(2) [Google Scholar]
  39. Mendez M.F. A neurology of the conservative-liberal dimension of political ideology. The Journal of Neuropsychiatry and Clinical Neurosciences. 2017;29(2):86–94. doi: 10.1176/appi.neuropsych.16030051. [DOI] [PubMed] [Google Scholar]
  40. Miller G.A., Buckhout R. Harper & Row; 1973. Psychology: The science of mental life. [Google Scholar]
  41. Miyagi M., Miyatani M., Nakao T. Relation between choice-induced preference change and depression. PLoS One. 2017;12(6):1–14. doi: 10.1371/journal.pone.0180041. [DOI] [PMC free article] [PubMed] [Google Scholar]
  42. Nakao T., Kanayama N., Katahira K., Odani M., Ito Y., Hirata Y.…Northoff G. Post response βγ power predicts the degree of choice-based learning in internally guided decision making. Scientific Reports. 2016;6(August):1–9. doi: 10.1038/srep32477. [DOI] [PMC free article] [PubMed] [Google Scholar]
  43. Nakao T., Ohira H., Northoff G. Distinction between externally vs. internally guided decision-making: Operational differences, meta-analytical comparisons and their theoretical implications. Frontiers in Neuroscience. 2012;6(31) doi: 10.3389/fnins.2012.00031. [DOI] [PMC free article] [PubMed] [Google Scholar]
  44. Nisbett R.E., Wilson T.D. Telling more than we can know: Verbal reports on mental processes. Psychological Review. 1977;84(3):231. [Google Scholar]
  45. Niv Y., Daniel R., Geana A., Gershman S.J., Leong Y.C., Radulescu A., Wilson R.C. Reinforcement learning in multidimensional environments relies on attention mechanisms. Journal of Neuroscience. 2015;35(21):8145–8157. doi: 10.1523/JNEUROSCI.2978-14.2015. [DOI] [PMC free article] [PubMed] [Google Scholar]
  46. Nosofsky R.M. The generalized context model: An exemplar model of classification. Formal Approaches in Categorization. 2011:18–39. doi: 10.1017/CBO9780511921322.002. [DOI] [Google Scholar]
  47. Pariser E. Penguin Group, The; 2011. The filter bubble: What the internet is hiding from you. [Google Scholar]
  48. Ramscar M., Yarlett D., Dye M., Denny K., Thorpe K. The effects of feature-label order and their implications for symbolic learning. Cognitive Science. 2010;34(6):909–957. doi: 10.1111/j.1551-6709.2009.01092.x. [DOI] [PubMed] [Google Scholar]
  49. Rescorla R.A., Wagner A.R. A theory of pavlovian conditioning: Variations in the effectiveness of reinforcement and nonreinforcement. Classical conditioning II: Current research and theory. 1972;2:64–99. [Google Scholar]
  50. Riefer P.S., Prior R., Blair N., Pavey G., Love B.C. Coherency-maximizing exploration in the supermarket. Nature Human Behaviour. 2017 doi: 10.1038/s41562-016-0017. [DOI] [PMC free article] [PubMed] [Google Scholar]
  51. Sauerland M., Sagana A., Siegmann K., Heiligers D., Merckelbach H., Jenkins R. These two are different. Yes, they’re the same: Choice blindness for facial identity. Consciousness and Cognition. 2016;40:93–104. doi: 10.1016/j.concog.2016.01.003. [DOI] [PubMed] [Google Scholar]
  52. Savage L.J. Wiley Publications in Statistics; 1954. The foundations of statistics. [Google Scholar]
  53. Schonberg, T., Bakkour, A., M Hover, A. Mumford, J. Nagar, L. Perez, J., & A Poldrack, R. (2014), Changing value through cued approach: An automatic mechanism of behavior change. Nature Neuroscience, 17. doi: 10.1038/nn.3673. [DOI] [PMC free article] [PubMed]
  54. Schultz W., Dayan P., Montague P.R. A neural substrate of prediction and reward. Science. 1997 doi: 10.1126/science.275.5306.1593. [DOI] [PubMed] [Google Scholar]
  55. Schulz, E., Bhui, R., Love, B. C., Brier, B., Todd, M. T., & Gershman, S. J. (2019). Structured, uncertainty-driven exploration in real-world consumer choice. Proceedings of the National Academy of Sciences, 201821028. doi: 10.1073/pnas.1821028116. [DOI] [PMC free article] [PubMed]
  56. Sharot T., De Martino B., Dolan R.J. How choice reveals and shapes expected hedonic outcome. Journal of Neuroscience. 2009 doi: 10.1523/JNEUROSCI.4972-08.2009. [DOI] [PMC free article] [PubMed] [Google Scholar]
  57. Sharot T., Velasquez C.M., Dolan R.J. Do decisions shape preference?: Evidence from blind choice. Psychological Science. 2010;21(9):1231–1235. doi: 10.1177/0956797610379235. 20679522 [DOI] [PMC free article] [PubMed] [Google Scholar]
  58. Strandberg T., Sivén D., Hall L., Johansson P., Pärnamets P. False beliefs and confabulation can lead to lasting changes in political attitudes. Journal of Experimental Psychology: General. 2018;147(9):1382–1399. doi: 10.1037/xge0000489. [DOI] [PubMed] [Google Scholar]
  59. Sutton R., Barto A. Reinforcement learning: An introduction. IEEE Transactions on Neural Networks. 1998;9(5):1054. doi: 10.1109/TNN.1998.712192. [DOI] [Google Scholar]
  60. Vinckier F., Rigoux L., Kurniawan I.T., Hu C., Bourgeois-Gironde S., Daunizeau J., Pessiglione M. Sour grapes and sweet victories: How actions shape preferences. PLoS Computational Biology. 2019;15(1):1–24. doi: 10.1371/journal.pcbi.1006499. [DOI] [PMC free article] [PubMed] [Google Scholar]
  61. Voigt K., Murawski C., Bode S. Endogenous formation of preferences: Choices systematically change willingness-to-pay for goods. Journal of Experimental Psychology: Learning, Memory, and Cognition. 2017;43(12):1872. doi: 10.1037/xlm0000415. [DOI] [PubMed] [Google Scholar]
  62. Voigt K., Murawski C., Speer S., Bode S. Hard decisions shape the neural coding of preferences. Journal of Neuroscience. 2019;39(4):718–726. doi: 10.1523/JNEUROSCI.1681-18.2018. [DOI] [PMC free article] [PubMed] [Google Scholar]
  63. von Neumann J., Morgenstern O., Rubinstein A. Princeton University Press; 1944. Theory of games and economic behavior (60th anniversary commemorative edition) [Google Scholar]
  64. Wegener D.T., Petty R.E., Detweiler-Bedell B.T., Jarvis W.B.G. Implications of attitude change theories for numerical anchoring: Anchor plausibility and the limits of anchor effectiveness. Journal of Experimental Social Psychology. 2001;37(1):62–69. doi: 10.1006/jesp.2000.1431. [DOI] [Google Scholar]
  65. Zajonc, R. B. (1968). Attitudinal effects of mere exposure. Journal of Personality and Social Psychology, 9 (2p2), 1 doi: 10.1037/h0025848. [DOI]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Supplementary material

mmc1.pdf (177.9KB, pdf)

RESOURCES