Skip to main content
Sage Choice logoLink to Sage Choice
. 2025 Oct 28;55(3):266–279. doi: 10.1177/03010066251387893

Sequential effects in facial attractiveness judgements: No evidence of stable individual differences

Robin S S Kramer 1,, Charlotte Cartledge 1
PMCID: PMC12979639  PMID: 41148247

Abstract

When items are judged in a sequence, evaluation of the current item is biased by the one preceding it. These sequential effects have been found for judgements of facial attractiveness, where studies have often shown an assimilation effect – ratings of the current face are pulled towards the attractiveness of the preceding face. However, the focus has been on the average bias across participants in general, with little consideration of individual differences. Here, we investigated an important first question – are individual differences in how sequential effects bias our judgements stable? Establishing this stability is crucial before considering potential associations between these individual differences in bias and other observer-level traits. To this end, we asked participants to provide attractiveness ratings for two different sequences of faces. In Experiment 1, one sequence comprised neutral, passport-style photos, while the other showed more unconstrained, naturalistic images. In Experiment 2, both sequences were composed of images taken from the same (constrained) photoset. Our results were identical for both experiments, with participants in general showing assimilation in their attractiveness judgements. However, for a given individual, we found no evidence that the strength of this bias was stable across the two sequences that they rated. These findings may be the result of within-person inconsistencies in perceiving facial attractiveness more broadly, and should serve to motivate further investigation of individual differences as applied to the domain of sequential effects.

Keywords: facial attractiveness, sequential effects, serial dependence, individual differences

Introduction

People are rarely perceived in isolation. As a result, the context in which they are viewed may influence our perceptions of them. For instance, judgements of attractiveness are higher when someone appears in a group (rather than alone; e.g., Kramer, Javorková, et al., 2024; Walker & Vul, 2014) or alongside an unattractive person (Kernis & Wheeler, 1981). Further, the temporal context also plays a role in the formation of first impressions. Sequential effects refer to biases caused by the previous item in a sequence when responding to the current item. In other words, the attractiveness of any given face depends, to some degree, on the attractiveness of the face that preceded it. While research has now established the presence of these biases in facial attractiveness judgements (e.g., Kramer et al., 2013; Xia et al., 2016), evidence is mixed as to the direction in which these biases act (e.g., Kondo et al., 2012; Pegors et al., 2015). In addition, since individuals vary in how much their judgements of the current face are influenced by the previous image (e.g., Huang et al., 2018), it is important to investigate whether these between-person differences are consistent/stable across tasks. If they are, we might begin to search for factors that account for these differences. However, few studies have considered this topic to date, and so here, we focus on the question of an individual's consistency in how sequential effects influence their facial attractiveness ratings.

Studies investigating sequential effects in attractiveness judgements have identified biases acting in opposing directions. Assimilation (or serial dependence; Manassi et al., 2023) refers to the current judgement being pulled towards the previous response (e.g., Kok et al., 2017; Kondo et al., 2012, 2013; Kramer et al., 2013; Taubert & Alais, 2016), whereas a contrast effect describes the current judgement being pushed away from the previous response (e.g., Huang et al., 2018; Pegors et al., 2015). Indeed, both types of bias may be present simultaneously while operating through different mechanisms (e.g., Pegors et al., 2015). A bias when responding to the current image may be caused by (1) the perception of the previous image's attractiveness (i.e., a perceptual or stimulus bias) and/or (2) the response given to the previous image (i.e., a response bias). Problematically, distinguishing between these two mechanisms can be difficult using traditional ratings tasks because of the strong correlation between the perceived attractiveness of the previous face and the rating it received. As a result, statistical artefacts due to multicollinearity can lead to uninterpretable analytical models (see Kramer & Jones, 2020).

To better distinguish between perceptual and response biases, researchers have utilised more complex experimental designs. For instance, Pegors et al. (2015) asked participants to alternate the type of judgement given on each trial (attractiveness and hair darkness). Their results demonstrated that the attractiveness rating given to the current face assimilated towards the hair darkness rating given to the previous face (a response bias) while contrasting away from the attractiveness value of the previous face (a perceptual bias). This pattern of results was replicated by Huang et al. (2018), who alternated the presentation of faces and ringtones. Here, no cross-modal contrast effect of the previous stimulus was found – the current face's attractiveness rating was not influenced by the agreeableness of the sound preceding it. However, assimilation due to the previous response remained – responding to the preceding ringtone biased the response given to the current face. Interestingly, the authors also demonstrated an assimilation effect, although weaker, when the previous and current responses were given orally, suggesting that response biases are unlikely to be caused by action repetition alone. This aligns with the findings of Kramer and Jones (Experiment 5; 2020), where relocating the mouse cursor to the centre of a circular scale before each response (and therefore minimising action repetition) failed to prevent assimilation from occurring.

Taking a different approach, Kramer and Pustelnik (2021) attempted to isolate each type of bias. In their first task, faces were presented in pairs, with the first viewed for 3 s (without being rated) before being replaced by the second face, which was rated for attractiveness. As such, a lack of response to the preceding face meant that any perceptual bias (here, there was evidence of a contrast effect) was isolated. In their second task, a rating was collected with no face presented (participants were simply instructed to respond with a specified value on the scale), and this was followed by a face that was rated for attractiveness. Therefore, responding to the current face without one preceding it meant that a response bias (which was absent in their findings) was isolated.

Another method for ruling out a response bias was developed by Xia and colleagues (2016). By requiring participants to rate the attractiveness of a sequence of faces twice, each time presented in a different random order, the researchers were able to use responses taken from the second (independent) run when modelling ratings given during the first run. These independent face ratings were not preceded by the same responses given during the first run, thereby removing the possibility of any response bias influencing the current image's rating during that first run. The findings of the study demonstrated an assimilation effect, where the attractiveness rating given to the current face was pulled towards the perceived attractiveness of the previous face (specifically, the difference in attractiveness between the previous and current faces).

Across all the studies mentioned so far, the focus has been on detecting and quantifying sequential effects across participants on average. In other words, do people in general show an assimilation or contrast effect when rating facial attractiveness? In a recent meta-analysis and review, Manassi and colleagues (2023) highlighted the study of individual differences as an important gap in the literature within the domain of sequential effects. For a given sample, the majority of participants may demonstrate assimilation towards the previous image in their judgements (for example) but there remain some who show no influence or even a contrast effect. However, these between-person differences have received little attention to date. In one study, the size of participant-level sequential effects during a facial identity task showed a small correlation with scores on a test of face recognition (Turbett et al., 2019). Importantly, to our knowledge, researchers have yet to investigate individual differences in relation to sequential biases during judgements of facial attractiveness.

Perceptions of attractiveness may be especially prone to individual differences in biases in comparison with other types of visual judgement. Low-level perceptual tasks tend to involve decisions based solely on the properties of the stimulus presented (e.g., brightness or size), and this may also be the case for the perception of faces when the task focusses on judging identity, for instance (e.g., Turbett et al., 2019). In contrast, facial attractiveness judgements are also strongly dependent on the individual perceiver, who is influenced by their own personal taste (e.g., Hönekopp, 2006; Kramer et al., 2018). Perhaps due to this subjectivity, judgements are often influenced by a variety of additional factors (e.g., presenting a face among others – Walker & Vul, 2014). Indeed, attractiveness judgements also vary for the same observer perceiving the same faces (e.g., Kramer et al., 2018, 2024). As such, given the somewhat unstable nature of these perceptions, we might predict that they would be particularly susceptible to sequential effects. Of course, this does not mean that an individual's bias resulting from the previous stimulus is stable across different sequences, and this provides the motivation for the investigation presented here.

In the current work, we therefore focussed on individual differences in sequential effects when rating facial attractiveness. Before researchers can explore whether such differences may be associated with other observer-level traits, we must first establish the stability of these individual differences. If the magnitude and direction of sequential effects are dictated by the individual then we should be able to detect a degree of consistency in these across similar versions of a task (e.g., the rating of attractiveness for two sequences of faces). However, if these effects are more temporary in nature (e.g., sensitive to the specific instance of the task or overshadowed by noise and other factors) then their magnitude/direction will not be stable across similar versions. Consequently, searching for additional factors associated with such biases would make little sense. While previous work has shown that individual differences in sequential effects in relation to orientation perception were highly stable when measured on two separate occasions (Kondo et al., 2022), this has yet to be considered within the domain of attractiveness judgements (or face perception more broadly). Here, across two experiments, we sought to address this question.

Experiment 1

In this first experiment, we investigated whether people showed consistency in how much their attractiveness ratings were influenced by sequential effects. By asking participants to provide ratings for two sets of face images taken from different databases, we were able to quantify this between-sequence consistency.

Method

Participants

A sample of 200 participants (136 women, 61 men, 3 nonbinary; age M = 29.2 years, SD = 14.9 years; 91% self-reported ethnicity as White) provided written informed consent online before taking part, and received an onscreen debriefing upon completion of the experiment. Participants were recruited by word of mouth (e.g., through asking friends and family, and sharing the experiment's weblink on social media) and were not paid to take part in the experiment. The data from 62 additional participants were excluded after failing to meet the predefined criteria (see below). Both Experiments 1 and 2 received ethical approval from the university's research ethics committee (ID 20105) and were carried out in accordance with the provisions of the World Medical Association Declaration of Helsinki.

The sample sizes for this experiment and Experiment 2 were determined a priori by conducting a power analysis using G*Power 3.1 (Faul et al., 2007). Since our main analysis would be a correlation, a total sample size of at least 138 participants was required to achieve 95% power to detect medium-sized effects at an alpha of .05 (two-tailed). In addition, the use of one-sample t-tests (following the same parameters) would require 54 participants. However, recruitment was allowed to continue until the end of a prespecified period.

Stimuli

For the first sequence, we randomly selected 30 images from a larger set of White women featured in the Chicago face database (CFD; Ma et al., 2015). In all cases, the women were displaying neutral facial expressions, with images containing the head and neck, as well as the top of the shoulders (see Figure 1). We chose to limit our set to a single gender and ethnicity to avoid additional influences on sequential effects that result from rating mixed sequences (Kramer et al., 2013).

Figure 1.

Figure 1.

The Two Styles of Image Presented in Experiment 1. Left: An image illustrative of those featured in the Chicago face database. Right: An image Iilustrative of the bridesmaids photoset. (The actual images used in the experiments cannot be reproduced here due to copyright restrictions.).

For the second sequence, we selected 30 images from a larger set featured in previous research which depicted White female bridesmaids collected through an online search engine (Carragher et al., 2021). Each image contained only the head and neck, with the face showing a positive expression (see Figure 1). Since these individuals were originally cropped from group photographs, the 30 images used here were chosen at random from the initial set with the proviso that no more than one image from each group photo was selected to avoid some images appearing more similar to others (e.g., by featuring a common background and dress style).

Procedure

The experiment was completed using the Gorilla online testing platform (Anwyl-Irvine et al., 2020). After consent was obtained, participants provided demographic information. Each participant was then presented with both sequences, twice each, in one of the following orders: ABAB or BABA. Order assignment was counterbalanced across participants.

For each sequence, participants were presented with the 30 face photos in a random order. For each image, participants were asked ‘how attractive is this face?’ and responded using a 7-point scale with labelled anchors (1 = low; 7 = high; see Figure 2). Images remained onscreen until ratings were provided by clicking the corresponding buttons. Responses were self-paced with no time limit. As soon as a rating was given, the next image appeared onscreen, with no intertrial interval.

Figure 2.

Figure 2.

An Example Trial During the Task, Featuring an Illustrative Image. (The actual images used in the experiments cannot be reproduced here due to copyright restrictions.).

Between each of the four blocks (two sequences × two repetitions), participants were presented with onscreen instructions where they were informed of their progress (i.e., how many blocks they had completed) and told to click the button provided when they were ready to continue.

Results

Exclusions

We identified and excluded participants who may not have attended to the task – a common concern when collecting online data (Hauser & Schwarz, 2016). First, if participants provided the same response for every image during one or more of the four blocks, their data were excluded (n = 8).

Second, since participants rated each sequence twice, we would expect these two sets of ratings for the same faces to be strongly correlated. As such, for each of the two sequences separately, we determined whether the participant's ratings were statistically different from chance/random responding. Following the approach of Xia and colleagues (2016), we carried out a permutation test. To produce the distribution of correlations expected from random responding, we permuted/shuffled the first set of ratings for a given sequence 1,000 times. For each permutation, we calculated the correlation between the two sets of ratings after this random shuffling of one set. The distribution of these 1,000 correlations was then used to calculate the significance level of the original correlation resulting from the unshuffled ratings (two-tailed; 0.05 threshold). This process was repeated for both sequences for each participant, where a nonsignificant permutation test for one or both sequences resulted in the exclusion of that participant (n = 54). Although relatively high, the proportion of participants excluded in this way (21%) was lower than the proportion of 29% reported by Xia and colleagues (2016).

Analytical Strategy

To quantify the influence of the previous image on the current rating, we followed the analytical strategy of Xia and colleagues (2016). 1 For each participant, for each of the two sequences separately, we fitted a multiple linear regression:

respi=β0+β1idpi+β2(idpi1idpi)

Here, resp i is the response given to the ith trial when the sequence was rated for the first time. As such, the preceding trial during that presentation is referred to as i−1. During the second viewing of that same sequence (now presented in a different, random order), the ratings given to these two images (trials i and i−1) were independent of each other (since i−1 no longer appeared before i) and are referred to as idp i and idp i −1 respectively. Finally, β0, β1 and β2 are the coefficients of the model.

To simplify, for a given image, the model predicts the first rating it received (i.e., when the sequence was first presented) using its second rating (i.e., received during its second presentation), along with the difference between the ratings given to it and the preceding image (but with both of these ratings taken from their second presentations). Since we used ratings from the second presentation as predictors in the model, any resulting sequential effects would be independent of a response bias. This is because a response bias is caused by responding to the previous image prior to the current one, while here, we utilised responses from the second presentation, meaning that idp i −1 was not given directly before idp i and, as a result, could not influence it.

In the model, β1 represents how well the second rating of an image predicts the first rating it received – presumably, this coefficient will be positive and large. Importantly, β2 represents the magnitude of any sequential effects. An attractiveness difference between the preceding face and the current one should produce assimilation (if the value is positive) or a contrast effect (if negative).

Regression Results

Across our sample of participants, β1 values were large (see Table 1). For both sequences, a one-unit increase in the second rating of attractiveness for a given face predicted an increase of close to 0.8 when that same face was first rated. As noted above, we expected a strong influence of this predictor since we were considering two responses to the same image. However, previous research has identified some within-person inconsistency when participants rated the same face twice (e.g., Kramer et al., 2018; Kramer, Ritchie, et al., 2024), explaining why the coefficient was not closer to a value of one here.

Table 1.

A Summary of the Regression Model Coefficients for Both Experiments.

Experiment Sequence β 1 β 2
1 CFD Set 1 0.77 [0.73, 0.81] 0.05 [0.03, 0.07]
Bridesmaids 0.79 [0.75, 0.83] 0.10 [0.07, 0.12]
2 CFD Set 1 0.72 [0.68, 0.75] 0.05 [0.03, 0.07]
CFD Set 2 0.77 [0.73, 0.80] 0.07 [0.05, 0.09]

Note. CFD = Chicago face database. Values are reported as means and 95% confidence intervals.

The β2 values in Table 1 quantified the influence of sequential effects for the two sequences. In both cases, positive values suggested that the response to the current face was pulled towards the attractiveness level of the previous face. One-sample t-tests demonstrated that these values were significantly larger than zero: CFD Set 1 – t(199) = 4.38, p < .001, Cohen's d = 0.31; Bridesmaids – t(199) = 7.86, p < .001, Cohen's d = 0.56. However, they remained small, with a one-unit increase in the difference between the (independent responses to the) current and previous faces predicting an increase of only 0.05 to 0.10 when that current face was first rated for attractiveness. Importantly, these values were comparable with the size of assimilation (β2 = 0.042) reported by Xia and colleagues (2016).

Finally, we correlated participants’ β2 values to determine whether there was evidence of a stable/consistent influence of sequential effects for the same person across the two sequences. We found a small, nonsignificant association: r(198) = .12, p = .106 (see Figure 3).

Figure 3.

Figure 3.

The Correlation Between β2 Values for the Two Sequences in Experiment 1.

Discussion

For both sequences, we found evidence of a small assimilation effect in attractiveness ratings resulting from a perceptual bias, in line with previous work (Xia et al., 2016). While this assimilation was present across the sample as a whole, individual differences in the magnitude and direction of any sequential effects were not stable across the two sequences. That is, participants who showed assimilation while rating the CFD images were not more likely to show this effect in their ratings of the bridesmaids’ images.

Both sequences involved rating the facial attractiveness of White women and so presumably incorporated similar cognitive and perceptual demands. However, the images in the two sequences differed somewhat in their style (i.e., standardised neutral photos versus unconstrained photos with positive expressions; see Figure 1). As such, the magnitude and direction of sequential effects evident in participants’ ratings may have been influenced by these characteristics of the image sets, which could have resulted in differing biases for the same participant. We therefore carried out a second experiment to address this possibility.

Experiment 2

The results of Experiment 1 showed no evidence of stable individual differences in the influence of sequential effects when rating two face sequences. However, differences in the image characteristics of the two sequences may have resulted in the absence of this stability. Therefore, in this second experiment, we again presented participants with two different sequences to rate for facial attractiveness. Crucially, both sequences comprised images taken from the same photoset (the CFD).

Method

Participants

A sample of 217 participants (152 women, 63 men, 1 nonbinary, 1 unreported; age M = 30.5 years, SD = 16.2 years; 94% self-reported ethnicity as White) provided written informed consent online before taking part, and received an onscreen debriefing upon completion of the experiment. Participants were recruited using the same approach as in Experiment 1. The data from 45 additional participants were excluded after failing to meet the predefined criteria (see below). There was no overlap between this sample and those who participated in Experiment 1.

Stimuli

For the first sequence, we used the first sequence from Experiment 1 – 30 images of White women taken from the CFD and displaying neutral facial expressions (Ma et al., 2015). For the second sequence, we randomly selected another 30 White women from the CFD, also displaying neutral facial expressions. As such, there was no overlap between the two sequences in terms of the identities featured.

Using the norming data provided alongside the CFD (Ma et al., 2015), we considered the attractiveness values of the faces in these two sequences: Sequence A – M = 3.35, SD = 0.89; Sequence B – M = 3.39, SD = 0.98. An independent samples t-test found no difference in their means, t(58) = 0.17, p = .864, Cohen's d = 0.04, while Levene's test found no difference in their variances, F(1, 58) = 0.46, p = .500.

Procedure

The procedure was identical to Experiment 1. Again, each participant was presented with both sequences, twice each, in one of the following orders: ABAB or BABA. As in Experiment 1, order assignment was counterbalanced across participants.

Results

Exclusions

We applied the same exclusion criteria as in Experiment 1. Here, we excluded four participants for providing the same response for every face during one or more of the four blocks. In addition, we excluded 41 participants because their ratings were not significantly different from random responding for one or both sequences.

Were the Two Sequences Perceived Similarly?

Since we aimed to investigate individual differences in sequential effects for two different sequences that were purposely chosen to be highly similar (i.e., two subsets of images taken from the same database), we began by confirming their similarity with respect to participants’ perceptions. First, ratings across all participants and images for the two sequences showed similar distributions (Set 1: M = 3.21, SD = 1.42; Set 2: M = 3.25, SD = 1.45; see Figure 4), mirroring the norming data provided by Ma and colleagues (2015) discussed earlier. Second, by correlating each participant's first and second ratings of the images for a given sequence, we found similarly high levels of within-person agreement for the two sequences (Set 1: M = 0.68, SD = 0.24; Set 2: M = 0.69, SD = 0.25; Fisher's r-to-z transformation and its inverse were applied as necessary). Third, between-person agreement (considering only participants’ first ratings) for the two sequences was also similarly high (Set 1: Cronbach's α = 0.96; Set 2: Cronbach's α = 0.96). Taken together, these descriptives suggested that the two sequences did not meaningfully differ in terms of the perception of attractiveness, although we acknowledge that low-level image properties were not explored (see the General Discussion).

Figure 4.

Figure 4.

The Distribution of Attractiveness Ratings for (a) CFD Set 1 and (b) CFD Set 2, collected across all participants and images. Frequencies shown here comprise first viewing’ ratings only.

Analytical Strategy

We followed the same analytical strategy as in Experiment 1.

Regression Results

Across our sample of participants, β1 values were large (see Table 1) and similar to those found in Experiment 1. Here, the β2 values were again positive and small (see Table 1), but significantly larger than zero: CFD Set 1 – t(216) = 4.88, p < .001, Cohen's d = 0.33; CFD Set 2 – t(216) = 6.45, p < .001, Cohen's d = 0.44. Finally, the correlation between participants’ β2 values was very small and nonsignificant: r(215) = .03, p = .694 (see Figure 5).

Figure 5.

Figure 5.

The Correlation Between β2 Values for the Two Sequences in Experiment 2.

Discussion

The results of this experiment mirrored those of Experiment 1. For both sequences, a small assimilation effect was present (Xia et al., 2016). However, individual differences in the magnitude and direction of any sequential effects were not stable across the two sequences despite both of these comprising images from the same photoset.

General Discussion

While numerous studies have investigated the influence of sequential effects on facial attractiveness judgements (e.g., Kondo et al., 2012; Kramer et al., 2013), researchers have yet to consider individual differences and their stability. Here, we carried out two experiments, both of which involved quantifying the magnitude of sequential effects when rating two sequences of faces. Participants, on average, showed an assimilation effect, where the rating given to the current image was pulled towards the perceived attractiveness of the previous face. While this assimilation was present in all sequences for the samples as a whole, individual differences across our participants were not stable when comparing the two sequences that they rated. In other words, demonstrating a strong assimilation effect while rating the first sequence (for example) was not an indicator that such an effect would influence ratings of the second sequence for any given individual.

In Experiment 1, we used sequences of images taken from two different photosets. In the CFD, images were tightly controlled/constrained, while in comparison, the photographs of bridesmaids were naturalistic in their style. Despite these differences, both sequences featured White women only and, as such, it seemed plausible that sequential effects might show some consistency across these two similar sequences requiring the rating of facial attractiveness. However, the lack of stable individual differences evident in our data might be explained by these photoset differences. In the CFD, only the identities varied but other aspects of the images (e.g., clothing, lighting, etc.) were held constant so sequential effects would have been the result of identity changes alone. For the bridesmaids, additional variability in clothing, facial expression, backgrounds and so on likely played a role in attractiveness judgements and, therefore, may have influenced the nature of how each image affected the next. Of course, the purpose of Experiment 2 was to rule out this possibility, and through the use of two sequences comprising images from the same photoset (the CFD), we again found no stability in how sequential effects influenced individuals’ ratings.

We acknowledge, however, that while the two sequences in Experiment 2 did not differ in terms of attractiveness (e.g., using previously collected norming data; Ma et al., 2015), low-level properties were not considered/controlled. Of course, two randomly selected subsets of images from the same photoset are also unlikely to substantially differ regarding low-level characteristics or other measures, and ultimately, any attempt to equate these two sequences for every conceivable dimension may be unachievable while still presenting two different sets of faces. Importantly, if the presence of any stability in individual differences were easily extinguished as a result of only minimal differences between the two sequences (e.g., in brightness or contrast levels) then this again highlights the lack of robust stability in any generalisable sense.

In the current work, we followed closely the design of Xia and colleagues (2016), allowing us to sidestep the complexities of potentially detecting, and therefore trying to differentiate between, perceptual and responses biases. By collecting ratings of each sequence twice, we were able to predict ratings given to the current item during the first presentation through the use of ratings taken from the second presentation, thus avoiding the responses biases that may have been present in the first run. While this facilitated a clearer understanding of (a lack of) stability in individual differences, we must acknowledge that our conclusions only inform regarding perceptual biases. We therefore recommend that future work should investigate whether response biases show stable individual differences, although a suitable experimental design is required before this question can be answered.

Here, we chose to limit our stimuli to White women only. Previous research has shown that unconstrained sequences with respect to gender and/or ethnicity led to local changes in sequential effects within the sequence itself (Kondo et al., 2013; Kramer et al., 2013). Since these additional influences were not the focus of the current work, we chose to avoid incorporating more variable sequences of this type. As such, our results may not generalise beyond the more limited sequences used here. However, we have no reason to predict that stability would be more likely for sequences of male or non-White faces. Indeed, for sequences incorporating a mix of genders and ethnicities, we might expect that stability across ratings of different sequences would be lower since this variation in sequential effects locally within each sequence should represent additional noise for any sequence-level measure. Of course, future study might provide further insight into this prediction.

While individual differences represent a much-neglected area of research within the domain of sequential effects (Manassi et al., 2023), one recent study demonstrated a high level of stability in biases with regard to orientation perception (Kondo et al., 2022). Further, Guan and Goettker (2024) found stable individual differences in the strength of sequential effects across two different low-level perceptual tasks (involving colour and orientation), but these were unrelated to a third, oculomotor task (tracking). However, assimilation of the orientation of a Gabor patch to the one preceding it, for instance, is likely to be a somewhat different instantiation of sequential effects processes to the high-level facial attractiveness judgements investigated here. Indeed, even for these low-level perceptions, stability in individual differences was limited to measures generated from presenting stimuli at the same visual field location (Kondo et al., 2022) or to tasks that were both perceptual in nature (Guan & Goettker, 2024). With such limitations even for low-level biases to demonstrate stability, it may be unsurprising that biases in high-level decision-making showed no evidence of stability in the current work.

Moving beyond sequential effects, individual differences in other biases when interpreting visual stimuli (e.g., apparent motion, structure from motion) appear to show stability when measured over multiple blocks of the same task, but also across different tasks (Wexler et al., 2022). However, the pattern of associations across tasks supports the idea that these individual differences may be stable but ‘local’ in the sense that they were associated for similar perceptual tasks only, for example, for two tasks involving two-dimensional motion but not across two- and three-dimensional motion tasks. Further complicating matters, Wexler and colleagues (2015) provided evidence that these individual differences in visual biases were stable but dynamic. That is, such ‘within observer’ biases were persistent over time while also showing gradual change (well-modelled by a ‘random walk’ approach). Of relevance to the current work, however, we would still expect associations between different measures of the same bias (here, the influence of sequential effects) within a given individual, even if such biases demonstrated this drift or change over time.

How might we explain the current findings when other face perception tasks have demonstrated stable individual differences more generally (e.g., Kramer et al., 2021), as have low-level perceptual tasks regarding sequential effects in particular (e.g., Kondo et al., 2022)? We propose that, for these high-level perceptions of facial attractiveness, the answer may be found in limited within-person consistency. Several studies have shown that associations between two sets of attractiveness ratings of the same image sequence, given by the same individual, were large but far from perfect (e.g., Kramer et al., 2018; Kramer, Ritchie, et al., 2024), and this result was replicated in the current work. Although these findings necessarily incorporated sequential effects, which would have served to lower the reported associations since each presentation of a sequence was in a different random order, the likelihood is that even with identical sequence presentations (i.e., images rated twice and presented in the same order each time), the correlation between ratings for a given individual would be notably lower than one. Therefore, this inconsistency in within-person judgements of attractiveness may serve to overshadow any stability in the influence of sequential effects (which are likely far smaller in magnitude). In contrast, low-level perceptions regarding the orientation of Gabor patches, for example, may be subject to less within-person instability, allowing for stable sequential effects to be detected (Kondo et al., 2022). Of course, this explanation may usefully form the basis for additional studies that could further our understanding of potential differences between high- and low-level perceptions and the relative strength of sequential effects.

To conclude, across two experiments, we have demonstrated a lack of stability in individual differences when quantifying the influence of sequential effects on facial attractiveness judgements. While our samples in general showed an assimilation effect, where ratings were pulled towards the attractiveness of the preceding face, we found no evidence of consistent within-person biases for ratings of two separate image sequences. To our knowledge, this study represents the first investigation of individual differences in reference to sequential effects while judging social traits from faces, and we strongly encourage further investigation within this domain.

Acknowledgments

The authors thank our Research Skills IV students for collecting the data.

1.

Note that we fitted a regression model separately for each participant’s data. In contrast, Xia and colleagues (2016) pooled the ratings from all participants before fitting a single model.

Footnotes

ORCID iD: Robin S. S. Kramer https://orcid.org/0000-0001-8339-8832

Author Contribution(s): Robin S. S. Kramer: Conceptualization; Data curation; Formal analysis; Investigation; Methodology; Project administration; Visualization; Writing – original draft; Writing – review & editing.

Charlotte Cartledge: Conceptualization; Investigation; Methodology; Writing – review & editing.

Funding: The authors received no financial support for the research, authorship, and/or publication of this article.

The authors declared no potential conflicts of interest with respect to the research, authorship, and/or publication of this article.

Data Availability: The data supporting both experiments are publicly available at: https://osf.io/843hk/

References

  1. Anwyl-Irvine A. L., Massonnié J., Flitton A., Kirkham N., Evershed J. K. (2020). Gorilla in our midst: An online behavioral experiment builder. Behavior Research Methods, 52, 388–407. 10.3758/s13428-019-01237-x [DOI] [PMC free article] [PubMed] [Google Scholar]
  2. Carragher D. J., Thomas N. A., Nicholls M. E. (2021). The dissociable influence of social context on judgements of facial attractiveness and trustworthiness. British Journal of Psychology, 112(4), 902–933. 10.1111/bjop.12501 [DOI] [PubMed] [Google Scholar]
  3. Faul F., Erdfelder E., Lang A. G., Buchner A. (2007). G*power 3: A flexible statistical power analysis program for the social, behavioral, and biomedical sciences. Behavior Research Methods, 39(2), 175–191. 10.3758/BF03193146 [DOI] [PubMed] [Google Scholar]
  4. Guan S., Goettker A. (2024). Individual differences reveal similarities in serial dependence effects across perceptual tasks, but not to oculomotor tasks. Journal of Vision, 24(12), 1–14. 10.1167/jov.24.12.2 [DOI] [PMC free article] [PubMed] [Google Scholar]
  5. Hauser D. J., Schwarz N. (2016). Attentive Turkers: MTurk participants perform better on online attention checks than do subject pool participants. Behavior Research Methods, 48(1), 400–407. 10.3758/s13428-015-0578-z [DOI] [PubMed] [Google Scholar]
  6. Hönekopp J. (2006). Once more: Is beauty in the eye of the beholder? Relative contributions of private and shared taste to judgments of facial attractiveness. Journal of Experimental Psychology: Human Perception and Performance, 32(2), 199–209. 10.1037/0096-1523.32.2.199 [DOI] [PubMed] [Google Scholar]
  7. Huang J., He X., Ma X., Ren Y., Zhao T., Zeng X., Li H., Chen Y. (2018). Sequential biases on subjective judgments: Evidence from face attractiveness and ringtone agreeableness judgment. PLOS ONE, 13(6), e0198723. 10.1371/journal.pone.0198723 [DOI] [PMC free article] [PubMed] [Google Scholar]
  8. Kernis M. H., Wheeler L. (1981). Beautiful friends and ugly strangers: Radiation and contrast effects in perceptions of same-sex pairs. Personality and Social Psychology Bulletin, 7(4), 617–620. 10.1177/014616728174017 [DOI] [Google Scholar]
  9. Kok R., Taubert J., Van der Burg E., Rhodes G., Alais D. (2017). Face familiarity promotes stable identity recognition: Exploring face perception using serial dependence. Royal Society Open Science, 4(3), 160685. 10.1098/rsos.160685 [DOI] [PMC free article] [PubMed] [Google Scholar]
  10. Kondo A., Murai Y., Whitney D. (2022). The test-retest reliability and spatial tuning of serial dependence in orientation perception. Journal of Vision, 22(4), 1–11. 10.1167/jov.22.4.5 [DOI] [PMC free article] [PubMed] [Google Scholar]
  11. Kondo A., Takahashi K., Watanabe K. (2012). Sequential effects in face-attractiveness judgments. Perception, 41(1), 43–49. 10.1068/p7116 [DOI] [PubMed] [Google Scholar]
  12. Kondo A., Takahashi K., Watanabe K. (2013). Influence of gender membership on sequential decisions of face attractiveness. Attention, Perception, & Psychophysics, 75(7), 1347–1352. 10.3758/s13414-013-0533-y [DOI] [PubMed] [Google Scholar]
  13. Kramer R. S. S., Mileva M., Ritchie K. L. (2018). Inter-rater agreement in trait judgements from faces. PLOS ONE, 13(8), e0202655. 10.1371/journal.pone.0202655 [DOI] [PMC free article] [PubMed] [Google Scholar]
  14. Kramer R. S. S., Javorková N., Jones A. L. (2024). No influence of face familiarity on the cheerleader effect. Visual Cognition, 32(3), 181–191. 10.1080/13506285.2024.2405700 [DOI] [Google Scholar]
  15. Kramer R. S. S., Jones A. L. (2020). Sequential effects in facial attractiveness judgments using cross-classified models: Investigating perceptual and response biases. Journal of Experimental Psychology: Human Perception and Performance, 46(12), 1476–1489. 10.1037/xhp0000869 [DOI] [PubMed] [Google Scholar]
  16. Kramer R. S. S., Jones A. L., Gous G. (2021). Individual differences in face and voice matching abilities: The relationship between accuracy and consistency. Applied Cognitive Psychology, 35(1), 192–202. 10.1002/acp.3754 [DOI] [Google Scholar]
  17. Kramer R. S. S., Jones A. L., Sharma D. (2013). Sequential effects in judgements of attractiveness: The influences of face race and sex. PLOS ONE, 8(12), e82226. 10.1371/journal.pone.0082226 [DOI] [PMC free article] [PubMed] [Google Scholar]
  18. Kramer R. S. S., Pustelnik L. R. (2021). Sequential effects in facial attractiveness judgments: Separating perceptual and response biases. Visual Cognition, 29(10), 679–688. 10.1080/13506285.2021.1995558 [DOI] [Google Scholar]
  19. Kramer R. S. S., Ritchie K. L., Flack T. R., Mireku M. O., Jones A. L. (2024). The psychometrics of rating facial attractiveness using different response scales. Perception, 53(9), 645–660. 10.1177/03010066241256221 [DOI] [PMC free article] [PubMed] [Google Scholar]
  20. Ma D. S., Correll J., Wittenbrink B. (2015). The Chicago face database: A free stimulus set of faces and norming data. Behavior Research Methods, 47, 1122–1135. 10.3758/s13428-014-0532-5 [DOI] [PubMed] [Google Scholar]
  21. Manassi M., Murai Y., Whitney D. (2023). Serial dependence in visual perception: A meta-analysis and review. Journal of Vision, 23(8), 1–29. 10.1167/jov.23.8.18 [DOI] [PMC free article] [PubMed] [Google Scholar]
  22. Pegors T. K., Mattar M. G., Bryan P. B., Epstein R. A. (2015). Simultaneous perceptual and response biases on sequential face attractiveness judgments. Journal of Experimental Psychology: General, 144(3), 664–673. 10.1037/xge0000069 [DOI] [PMC free article] [PubMed] [Google Scholar]
  23. Taubert J., Alais D. (2016). Serial dependence in face attractiveness judgements tolerates rotations around the yaw axis but not the roll axis. Visual Cognition, 24(2), 103–114. 10.1080/13506285.2016.1196803 [DOI] [Google Scholar]
  24. Turbett K., Palermo R., Bell J., Burton J., Jeffery L. (2019). Individual differences in serial dependence of facial identity are associated with face recognition abilities. Scientific Reports, 9, 18020. 10.1038/s41598-019-53282-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
  25. Walker D., Vul E. (2014). Hierarchical encoding makes individuals in a group seem more attractive. Psychological Science, 25(1), 230–235. 10.1177/0956797613497969 [DOI] [PubMed] [Google Scholar]
  26. Wexler M., Duyck M., Mamassian P. (2015). Persistent states in vision break universality and time invariance. Proceedings of the National Academy of Sciences, 112(48), 14990–14995. 10.1073/pnas.1508847112 [DOI] [PMC free article] [PubMed] [Google Scholar]
  27. Wexler M., Mamassian P., Schütz A. C. (2022). Structure of visual biases revealed by individual differences. Vision Research, 195, 108014. 10.1016/j.visres.2022.108014 [DOI] [PubMed] [Google Scholar]
  28. Xia Y., Leib A. Y., Whitney D. (2016). Serial dependence in the perception of attractiveness. Journal of Vision, 16(15), 28–28. 10.1167/16.15.28 [DOI] [PMC free article] [PubMed] [Google Scholar]

Articles from Perception are provided here courtesy of SAGE Publications

RESOURCES