Skip to main content
Frontiers in Psychology logoLink to Frontiers in Psychology
. 2023 Jan 26;14:1029773. doi: 10.3389/fpsyg.2023.1029773

Happy and sad music acutely modulate different types of attention in older adults

Nicholas Dovorany 1, Schea Brannick 2, Nathan Johnson 1, Ileana Ratiu 2,3, Arianna N LaCroix 4,*
PMCID: PMC9909555  PMID: 36777231

Abstract

Of the three subtypes of attention outlined by the attentional subsystems model, alerting (vigilance or arousal needed for task completion) and executive control (the ability to inhibit distracting information while completing a goal) are susceptible to age-related decline, while orienting remains relatively stable. Yet, few studies have investigated strategies that may acutely maintain or promote attention in typically aging older adults. Music listening may be one potential strategy for attentional maintenance as past research shows that listening to happy music characterized by a fast tempo and major mode increases cognitive task performance, likely by increasing cognitive arousal. The present study sought to investigate whether listening to happy music (fast tempo, major mode) impacts alerting, orienting, and executive control attention in 57 middle and older-aged adults (M = 61.09 years, SD = 7.16). Participants completed the Attention Network Test (ANT) before and after listening to music rated as happy or sad (slow tempo, minor mode), or no music (i.e., silence) for 10 min. Our results demonstrate that happy music increased alerting attention, particularly when relevant and irrelevant information conflicted within a trial. Contrary to what was predicted, sad music modulated executive control performance. Overall, our findings indicate that music written in the major mode with a fast tempo (happy) and minor mode with a slow tempo (sad) modulate different aspects of attention in the short-term.

Keywords: attention, aging, music, emotion, ant, alerting, executive control, orienting

1. Introduction

Attention is a multifaceted construct that moderates our interactions with the world (Petersen and Posner, 2012; Harvey, 2019). It also supports higher order cognitive processes such as working memory and executive functions (McCabe et al., 2010). There are several models that divide attention into distinct components (Sohlberg and Mateer, 1987; Mirsky et al., 1991), with one prominent model dividing attention into three subsystems: alerting, orienting, and executive control (Posner and Petersen, 1990). Alerting involves vigilance to perceive information, orienting is the ability to prioritize a specific aspect or location of a stimulus, and executive control is the ability to complete goal directed behavior while inhibiting distracting information (Posner and Petersen, 1990).

The Attention Network Test (ANT) simultaneously assesses alerting, orienting, and executive control using a cued-flanker task (Fan et al., 2002). In this computerized task, participants complete a flanker task (i.e., determine which direction a center arrow, in a series of five arrows, is pointing). A congruent trial occurs when all the arrows point in the same direction, an incongruent trial occurs when the center arrow points in the opposite direction of the flanking arrows, and a neutral trial occurs when the center arrow is presented with no flanking arrows. Participants are either cued to the onset of the flanker task with an alerting cue, the location of the flanker task with an orienting cue or provided with no cueing about when or where the upcoming flanker task will be presented. The efficiency of each attention subsystem is measured by comparing specific combinations of cues and flanker targets (described below in “attention network task”).

The ANT was initially developed and tested in younger adults (Fan et al., 2002), however it has since been used to test all three attentional subsystems in a variety of populations, including older adults (Gamboz et al., 2010; Zhou et al., 2011; Knight and Mather, 2013). Across several studies, older adults appear to be less alert, or benefit less from the alerting cue, than their younger counterparts (Jennings et al., 2007; Gamboz et al., 2010; Zhou et al., 2011; Knight and Mather, 2013; Kaufman et al., 2016; Williams et al., 2016). Similarly, for executive control, research indicates that executive control declines with increasing age. Compared to younger adults, older adults have greater difficulty resolving incongruent trials that contain conflict compared to neutral trials that are without conflict (Jennings et al., 2007; Gamboz et al., 2010; Mahoney et al., 2010; Zhou et al., 2011; Lu et al., 2016). In contrast to alerting and executive control, the ability to orient attention remains relatively stable with increasing age (Fernandez-Duque and Black, 2006; Jennings et al., 2007; Gamboz et al., 2010; Mahoney et al., 2010; Zhou et al., 2011). Alerting and executive control may be particularly susceptible to age-related decline, compared to orienting attention, because they are supported by the frontal lobes (among other regions; Roberts and Hall, 2008; Anderson et al., 2010; Petersen and Posner, 2012; Rinne et al., 2013). The frontal lobes atrophy as part of the normal aging process (Jernigan et al., 2001; Peters, 2006; Glisky, 2007; Zanto and Gazzaley, 2019). For instance, research shows that prefrontal cortical volume declines at a rate of approximately 5% per decade starting around the age of 40 (Svennerholm et al., 1997; Jernigan et al., 2001; Walhovd et al., 2005; Peters, 2006; Long et al., 2012). In contrast, more posterior regions, including regions associated with orienting attention in the parietal lobes (Corbetta and Shulman, 2002; Petersen and Posner, 2012), are less susceptible to cortical volume loss as a function of age (Trollor and Valenzuela, 2001; Scahill et al., 2003; Peters, 2006).

There are currently few strategies that have been shown to maintain attentional abilities in older adults. This is problematic as older adults report decreased quality of life (Hoe et al., 2009; Williams and Kemper, 2010) in concert with symptoms of broadly reduced attention and cognition. Listening to music may be one mechanism to help older adults maintain cognition, including attentional performance (Koelsch, 2010, 2014). In an early study, Rauscher et al. (1993) instructed participants to listen to Mozart’s sonata, an instructive relaxation tape, or silence for 10 min before completing a visuospatial task. Their results showed that participants who listened to Mozart performed better on the visuospatial task than the participants in the other two conditions. This effect, which describes better cognitive performance after listening to classical music (Rauscher et al., 1993), was popularized by the media and coined the “Mozart effect.” While Rauscher et al. (1995) replicated their original results in a follow up study, subsequent studies conducted by other research groups, failed to replicate this so called “Mozart effect” (Steele et al., 1997, 1999; McCutcheon, 2000), which led to the development of a new theory: the “arousal-mood” hypothesis (Husain et al., 2002).

The “arousal-mood” hypothesis posits that the tempo and composition mode of music increases arousal and mood, which subsequently increases cognitive performance (Husain et al., 2002). This hypothesis arose in part from research showing that music composed with a fast tempo and in the major mode increases arousal and feelings of happiness, while musical pieces with a slow tempo written in the minor mode decreases arousal and induces feelings of sadness (Balch et al., 1999; Gabrielsson and Lindström, 2001; Thompson et al., 2001; Husain et al., 2002; McConnell and Shore, 2011b). The “arousal-mood” hypothesis was first tested by having participants listen to one of three acoustic conditions: Mozart’s sonata (a happy-sounding, composition with a fast tempo predicted to increase arousal and induce feelings of happiness), Albinoni adagio (a sad-sounding, slow piece predicted to decrease participant arousal and induce feelings of sadness), or 10 min of silence (Thompson et al., 2001). Participants then completed the same visuospatial task used by Rauscher et al. (1993). Arousal level and mood were measured using questionnaires before and after music listening. Thompson et al. (2001) found that participants who listened to Mozart performed better on the visuospatial task than those who listened to Albinoni or silence. The Mozart group also reported increased arousal and happiness compared to the Albinoni group. The authors argued that participants’ cognitive performance improved because listening to Mozart’s sonata increased arousal and induced feelings of happiness (Thompson et al., 2001). Subsequent studies testing this theory have drawn similar conclusions: listening to happy music, written in the major mode with a fast tempo, enhances performance on a variety of visuospatial and processing speed tasks (Thompson et al., 2001; Husain et al., 2002; Schellenberg and Hallam, 2005; Schellenberg et al., 2007; Schellenberg, 2012), working memory tasks (Mammarella et al., 2007; Chew et al., 2016; Palmiero et al., 2016), and attention tasks (Marti-Marca et al., 2020; Kiss and Linnell, 2021), including investigations of executive control (Finucane et al., 2010; Bruyneel et al., 2013; Fernandez et al., 2019). It is possible that the fast tempo and major mode of happy music increases attention, particularly executive control by not only increasing arousal and mood (Balch et al., 1999; Gabrielsson and Lindström, 2001; Thompson et al., 2001; Husain et al., 2002), but also by engaging the brain regions which support attention (Fernandez et al., 2019). Happy music may increase alerting attention similarly to executive control since both subsystems are associated with similar regions in the frontal cortex (Fan et al., 2005; Roberts and Hall, 2008; Anderson et al., 2010; Petersen and Posner, 2012; Rinne et al., 2013).

The purpose of this study was to investigate how listening to happy music, written in the major mode with a fast tempo, impacts alerting, orienting, and executive control attention in middle and older-aged adults. We specifically recruited non-musicians to test for acute changes in attention, as such effects in musicians would be less detectable because long-term musical training preserves cognition, including executive control (Román-Caballero et al., 2018; Medina and Barraza, 2019). The attentional subsystems were assessed using the ANT before (time one) and after (time two) participants listened to happy music (major mode, fast tempo), sad music (minor mode, slow tempo), or no music (i.e., silence) for 10 min. Participants listened to music prior to completing the task as the irrelevant sound effect consistently demonstrates that listening to background music impairs task performance (Ransdell and Gilroy, 2001; Cassidy and MacDonald, 2007; Rowe et al., 2007; Jiang et al., 2011; Rey, 2012; Gonzalez and Aiello, 2019; Cloutier et al., 2020). We hypothesized that listening to happy music, but not sad or no music, would increase alerting and executive control from time one to time two. No changes in orienting attention were expected from time one to time two for any group. Alerting and executive control are known to interact (Callejas et al., 2004; Fan et al., 2009; Gamboz et al., 2010; Ishigami and Klein, 2010; McConnell and Shore, 2011a; Weinbach and Henik, 2012). Therefore, we expected a time x congruency x cue x group interaction such that the alerting effect within incongruent trials would not be present at time one for any group and would only be present at time two for the happy music listening group. We did not expect happy music to impact congruent or neutral trials since the alerting effect should be present at time one for both trial types.

2. Materials and methods

2.1. Participants

Fifty-nine neurotypical adults were recruited for this study. Two participants were excluded for failing to follow instructions (n = 2), resulting in 57 participants (48 females) who ranged in age from 50 to 84 years (M = 61.09, SD = 7.16) being included in the final sample. All participants spoke English and had no self-reported history of psychiatric or neurological disease and reported normal or corrected to normal vision. Participants were assigned to one of three experimental groups at the time of study enrollment. The three groups did not differ by age, gender, education, hearing (pure tone average of 500–4,000 Hz in both ears), years of musical training, testing time, basic cognitive abilities (assessed via the Montreal Cognitive Assessment MoCA; Nasreddine et al., 2005), caffeine intake, sleep, or exercise (Table 1). Participants were monetarily compensated for their time. Midwestern University’s Institutional Review Board approved all procedures, and all participants provided informed consent for the study.

Table 1.

Means and standard deviations (in parentheses) for the demographic and study specific comparisons between the happy (n = 19), sad (n = 20), and no music (n = 18) listening groups.

Happy music Sad music No music Statistic
Age 62.37 (8.06) 59.70 (7.01) 61.28 (6.39) F (2,54) = 0.68, p = 0.51
Gender 0.84 (0.38) 0.85 (0.37) 0.89 (0.32) F (2,54) = 0.01, p = 0.99
Education (years) 16.84 (4.10) 15.90 (2.57) 15.56 (4.13) F (2,54) = 0.62, p = 0.54
MoCA 26.26 (2.02) 26.90 (2.38) 26.11 (1.64) F (2,54) = 0.81, p = 0.45
Pure tone average 19.38 (7.36) 16.22 (7.88) 20.14 (8.10) F (2,54) = 1.40, p = 0.26
Musical training (years) 1.74 (2.88) 1.35 (2.66) 5.56 (14.06) F (2,54) = 1.48, p = 0.24
Participation time (military) 1197.63 (246.10) 1314.75 (252.26) 1191.11 (276.72) F (2,54) = 1.41, p = 0.25
Sleep (hours) 7.24 (1.10) 6.73 (1.01) 7.19 (0.86) F (2,54) = 1.59, p = 0.21
Caffeine (ounces) 7.32 (0.81) 9.00 (6.34) 5.56 (7.66) F (2,54) =1.03, p = 0.36
Exercise (minutes) 10.26 (19.89) 6.00 (12.31) 10.00 (30.87) F (2,54) = 0.23, p = 0.80

2.2. Experimental procedures

Participants were assigned to one of three experimental groups upon study enrollment. Group assignment was initially randomized. Toward the end of data collection, a minimization approach was adopted to ensure the groups were comparable on variables of interest such as age, gender, and testing time (Saint-Mont, 2015). Two groups required participants to listen to music, happy music (n = 19) or sad music (n = 20), and the third group was a no music listening (or silent) control condition (n = 18). All groups completed the same study procedures (see Figure 1). Participants’ attention was measured using the Attention Network Test (ANT) before (ANT-T1) and after (ANT-T2) listening to music or silence for 10 min. The ANT and listening tasks were programmed in E-prime 3.0 (Psychology Software Tools, Pittsburgh, PA) on an EyeLink 1,000 Plus eye-tracker (SR Research Ltd., Ottawa, ON, Canada). These additional physiological measures of attentional effort were collected as part of a larger project aiming to answer complementary, yet different questions about the time-course of attention following music listening; as such, analyzes of these data are beyond the scope of this paper.

Figure 1.

Figure 1

Visual representation of the experimental design.

Participants’ cognitive status and hearing were screened using the MoCA and pure tone averages, respectively. Participants additionally completed a questionnaire which asked about demographic information, music familiarity, caffeine intake, sleep, exercise, and general health. The NASA Task Load Index (TLX; Hart and Staveland, 1988) was administered at three timepoints throughout the experiment to assess self-perceived effort after the ANT-T1 (NASA-T1), after listening to music or silence for 10 min (NASA-T2), and after completing the ANT-T2 (NASA-T3).

2.3. Music stimuli

Two non-lyrical, classical music pieces were selected based on work done by Trost et al. (2012). These authors had participants rate various musical pieces based on evoked emotions using the Geneva Emotional Music Scale (Zentner et al., 2008), and arousal and valence using a series of multidimensional questionnaires on adjectives of emotion. The “happy music” piece we selected was “Violin Sonata in F Major, 3. Assai vivace” by Felix Mendelssohn. This piece is written in the major mode with a fast tempo and is associated with joyfulness, as well as feelings of high arousal and high valence (Trost et al., 2012). The “sad music” piece we selected was “String Quartet, No. 8 in C Minor, Op. 110, 1. Largo” by Dimitri Shostakovich. This piece is written in the minor mode with a slow tempo and has been associated with sadness and low arousal and low valence (Trost et al., 2012). In addition to evoking opposing emotions, these two compositions were equally familiar to participants (Trost et al., 2012). The original duration of the happy and sad music pieces was 5:14 and 4:53 min, respectively. Each musical piece was presented on a loop through noise canceling headphones for a total of 10 min at a sound level of 60 decibels. The no music listening groups’ (or control groups’) listening procedure was identical to the happy and sad music listening groups’, except they listened to silence for 10 min rather than music. For the two music listening groups, participants rated their perceived emotion following music listening using a visual analog scale anchored at “0″ (sadness) and “100″ (happiness); a rating of “50″ indicated neutral emotion. Participants also rated their familiarity with the musical composition they listened to using a questionnaire which asked if they were unfamiliar (0), somewhat familiar (1), or very familiar (2) with the music.

2.4. Attention network task

Participants completed a modified version of the ANT developed by Fan et al., 2002 (Figure 2). Each trial began with a fixation cross for 2,400 milliseconds. Following the offset of the fixation cross, a visual cue was presented for 100 milliseconds. Visual cue conditions were as follows: (1) spatial cue (single asterisk presented either above or below the fixation cross; spatial cues predicted the location of the flanker task 75% of the time (Fan et al., 2009); these will be referred to as valid spatial cues), (2) double cues (simultaneous presentation of one asterisk above the fixation cross and one asterisk below the fixation cross), (3) center cue (single asterisk presented in the center of the screen), and (4) no cue (no cueing was provided; i.e., no offset of the fixation cross). Following the offset of the cue, the fixation cross was presented on the screen for 400 milliseconds after which time participants were presented with the flanker stimuli either above or below the fixation cross.

Figure 2.

Figure 2

The Attention Network Test (ANT) procedure, cues, and targets.

The flanker stimuli consisted of five arrows pointing left or right. A congruent trial occurred when the center arrow was flanked by arrows pointing in the same direction (i.e., all arrows pointing to the left or right). An incongruent trial occurred when the center arrow pointed in the opposite direction of the flanking arrows (e.g., the center arrow points left, and the flanking arrows point right). A neutral trial occurred when the center arrow was flanked by dashed lines (two on each side). Accuracy and reaction time were collected via keyboard button press. Participants were instructed to be fast and accurate and press the left arrow key if the center arrow pointed to the left and the right arrow key if the center arrow pointed to the right. Participants completed a total of 156 trials across two blocks (78 trials x 2 blocks = 156 total trials) at each timepoint (ANT-T1, ANT-T2). For each ANT, all cues were presented equally (36 trials) across both blocks, except for the invalid spatial cue which was presented on 12 trials across both blocks. All flanker targets were presented equally (52 trials per condition) across both ANT blocks. This combination of cues and targets allowed us to calculate three measures of attention using this single task: alerting (no cue-double cued trials), orienting (center cued-valid spatially cued trials), and executive control (incongruent-neutral trials). These difference scores are interpreted as follows: large, positive difference scores equal better alerting and orienting and difference scores closer to zero equal better executive control. Trial presentation was randomized for each participant. Verbal and written instructions, examples of all stimuli, and 10 practice trials preceded the start of the experiment.

2.5. NASA task load index (NASA-TLX)

The NASA-TLX (Hart and Staveland, 1988) was used to measure participant’s perceived mental workload, attentional allocation, and physical effort at three timepoints throughout the experiment (NASA-T1, NASA-T2, NASA-T3; Figure 1). The NASA-TLX consists of six questions and was administered as originally published. Participants rated their perceived effort using a visual analog scale anchored at “0” (very low effort) and “100” (very high effort). Participant’s average raw score across all six questions was analyzed to ensure self-perceived effort was equivalent across groups as several studies indicate that effort affects cognitive task performance (Inzlicht et al., 2018).

2.6. Statistical analysis

2.6.1. Perceptions of music listening condition

All data were analyzed in SPSS Version 27 (IBM, 2020). Independent samples t-tests were used to compare how participants in the happy and sad music listening groups rated the emotions evoked by each composition and their familiarity with the composition. Changes in participants self-reported effort across the experiment using the NASA-TLX were analyzed using a mixed ANOVA with three levels of group (happy music, sad music, no music listening control group) and three levels of time (NASA-T1, NASA-T2, and NASA-T3); the dependent variable was the average raw score across all six questions.

2.6.2. Changes in attention after music listening

Accuracy and reaction time (RT; measured in milliseconds) were collected for each ANT trial. Reaction time is our behavioral measure of interest as it is widely accepted that older adults demonstrate a speed-accuracy tradeoff where they favor accuracy over speed (see Heitz, 2014 for a review). Only reaction times associated with correct responses were included in the analyzes (0.6% of trials were removed for incorrect responses). We first used paired sample t-tests to test our hypothesis that listening to happy music would increase alerting and executive control attention from time one to time two, but sad and no music listening would not. We used reaction time difference scores to calculate the alerting (no cue–double cue), orienting (center cue–valid spatial cue), and executive control (incongruent–neutral trials) effects at time one and time two for each group. Reaction time difference scores allowed us to isolate any music-induced changes to alerting, orienting, and executive control from individual variability in general processing speed, motor response speeds, and practice effects.

A 3×2×3×4 mixed ANOVA with three levels of group (happy music, sad music, no music listening control group), two levels of time (ANT-T1, ANT-T2), three levels of congruency (congruent, incongruent, neutral), and four cue types (no cue, double cue, center cue, and valid spatial cue1) was conducted to test our hypothesis that group, time, cue, and congruency would interact such that happy music would change the efficiency of the alerting subsystem from time one to time two within incongruent trials only. The Greenhouse–Geisser correction was used to correct for violations to sphericity. Post-hoc comparisons were corrected for multiple comparisons using the Benjamini-Hochberg (BH) false discovery rate (FDR; Benjamini and Hochberg, 1995).

3. Results

3.1. Perceptions of music listening condition

Participants’ familiarity with each musical composition did not differ [t(37) = 1.13, p = 0.27; Happy Music: M = 0.79, SD = 0.71; Sad Music: M = 0.55, SD = 0.61]. Consistent with expectations, there was a significant difference in how participants rated the emotions tied to each musical piece [t(37) = 6.39, p < 0.001]: participants rated the happy music (M = 79.74, SD = 12.19) higher on the visual analog scale than the sad music (M = 44.75, SD = 20.68). While there was no effect of group on the NASA-TLX [F(2, 53) = 0.45, p = 0.64] and group did not interact with time [F(4, 106) = 0.87, p = 0.49], all participants rated the music listening phase as the most cognitively taxing [F(1.46, 77.24) = 22.74, p < 0.001; NASA-T1: M = 15.94, SD = 12.39; NASA-T2: M = 27.54, SD = 19.44; NASA-T3: M = 19.63, SD = 15.39].

3.2. Changes in attention after music listening

3.2.1. Paired sample t-tests

The means and standard deviations for the ANT separated by group and time are reported in Table 2. Reaction time difference scores for each group on the ANT-T1 (time one) and the ANT-T2 (time two) are plotted in Figure 3. Individual variability in how alerting, orienting, and executive control change from time one to time two are depicted in Figure 4. As expected, the happy music listening group was more alert at time two than time one [i.e., they had larger reaction time difference scores at time two compared to time one; t(18) = 2.09, p = 0.05, d = 0.48]. Alertness did not change from time one to time two for the sad [t(19) = 0.86, p = 0.40, d = 0.19] or no music listening groups [t(17) = 0.59, p = 0.56, d = 0.14]. The orienting effect did not change from time one to time two for any group: happy [t(18) = 0.33, p = 0.74, d = 0.08], sad [t(19) = 0.31, p = 0.76, d = 0.07], and no music listening [t(17) = 0.91, p = 0.38, d = 0.21]. Unexpectedly, executive control increased from time one to time two for the sad music listening group [i.e., they had smaller reaction time difference scores at time two than time one; t(19) = 2.07, p = 0.05, d = 0.46]. Executive control did not change from time one to time two for the happy [t(18) = 0.32, p = 0.75, d = 0.07] or no music listening groups [t(17) = 0.99, p = 0.34, d = 0.23].

Table 2.

The means and standard deviations (in parentheses) for the ANT separated by group and time.

Cue types Congruency conditions
Group No cue Double Center Valid spatial Invalid spatial Congruent Incongruent Neutral All
ANT-T1
Happy 702.42 (116.55) 671.46 (98.47) 683.82 (104.47) 651.11 (102.31) 708.26 (117.03) 644.24 (95.30) 761.77 (118.55) 633.27 (102.16) 679.76 (103.7)
Sad 657.39 (100.97) 621.58 (92.84) 633.26 (91.95) 615.83 (96.26) 675.36 (118.20) 604.08 (97.35) 714.49 (98.19) 588.21 (91.09) 635.59 (93.94)
Control 673.66 (102.72) 660.90 (101.76) 661.84 (105.17) 634.17 (87.99) 696.87 (140.65) 622.64 (97.60) 742.80 (112.74) 617.65 (95.65) 661.03 (100.32)
All 677.54 (106.71) 650.63 (98.33) 659.14 (100.91) 633.38 (95.27) 693.12 (123.87) 623.33 (96.47) 739.19 (109.74) 612.53 (96.50) 658.35 (99.24)
ANT-T2
Happy 667.68 (89.37) 612.56 (79.80) 636.93 (87.95) 609.59 (104.24) 642.71 (81.36) 596.08 (82.24) 716.89 (90.59) 585.77 (86.41) 632.91 (83.42)
Sad 638.68 (107.50) 596.11 (94.72) 596.43 (90.35) 582.04 (96.38) 627.55 (98.66) 571.61 (96.10) 677.90 (100.77) 566.72 (92.98) 605.41 (94.68)
Control 647.98 (101.50) 628.70 (110.91) 620.09 (98.83) 582.93 (79.23) 636.40 (88.15) 590.79 (94.58) 707.46 (112.70) 567.96 (82.44) 622.07 (93.48)
All 651.29 (98.82) 611.89 (94.89) 617.40 (92.25) 591.51 (93.38) 635.39 (88.49) 585.82 (90.21) 700.23 (101.13) 573.46 (86.45) 619.84 (89.8)
ANT: collapsed across time
Happy 685.05 (99.37) 642.01 (87.56) 660.38 (92.00) 630.35 (99.18) 675.49 (91.72) 620.16 (84.16) 739.33 (103.27) 609.52 (91.17) 656.34 (91.68)
Sad 648.04 (97.95) 608.85 (90.32) 614.84 (86.68) 598.94 (92.25) 651.45 (97.76) 587.84 (90.70) 696.19 (95.54) 577.46 (88.25) 620.50 (90.27)
Control 660.82 (95.04) 644.80 (96.72) 640.97 (93.85) 608.55 (76.44) 666.63 (105.99) 606.72 (88.72) 725.13 (103.38) 592.81 (81.43) 641.55 (89.60)
All 664.41 (97.03) 631.26 (91.38) 638.27 (91.15) 612.44 (89.45) 664.26 (97.27) 604.58 (87.41) 719.71 (100.53) 592.99 (86.64) 639.09 (90.16)
Figure 3.

Figure 3

Changes in alerting (A) reaction time difference score for no cue–double cue, orienting (B) reaction time difference score for center cue–valid spatial cue, and executive control (C) reaction time difference score for incongruent–neutral trials from time one (T1) to time two (T2) within each group. Error bars represent ± 1 standard error. Alerting improved from time one to time two for the happy music group only. Executive control improved from time one to time two for the sad music group only.

Figure 4.

Figure 4

Individual variation in how alerting (reaction time difference score for no cue–double cue), orienting (reaction time difference score for center cue–valid spatial cue), and executive control (reaction time difference score for incongruent–neutral trials) change from time one to time two for each participant. Green shading represents better attention and red shading poorer attention. The individual data depicts a similar pattern as the whole-group analysis: alerting attention predominately changes in response to happy music and executive control in response to sad music.

3.2.2. Mixed ANOVA

The means and standard deviations for the ANT separated by time, group, congruency, and cue are reported in Table 3. The full mixed ANOVA results are reported in Table 4. There were no group differences between the happy, sad, and no music listening groups [F(2, 54) = 0.81, p = 0.45]. The main effect of time was significant [F(1, 54) = 23.71, p < 0.001]: as expected, all participants produced faster responses at time two than time one. The main effect of cue was significant [F(2.61, 140.78) = 57.45, p < 0.001]: both the alerting (faster responses on double cue than no cue trials) and orienting effects were significant (faster responses on valid spatial cue than center cue trials). The main effect of congruency was also significant [F(1.60, 86.63) = 610.66, p < 0.001]: participants responded more slowly on incongruent trials than neutral trials (Tables 2, 4). The congruency x cue interaction was significant [F(4.81, 259.87) = 8.02, p < 0.001]: the alerting effect was significant for congruent and neutral trials, but not the incongruent trials. The orienting effect was significant in all three congruency conditions (Table 4).

Table 3.

The means and standard deviations (in parentheses) for the ANT separated by time, group, congruency, and cue.

Congruent trials Incongruent trials Neutral trials
Group No Cue Double Center Valid spatial No cue Double Center Valid spatial No Cue Double Center Valid spatial
ANT-T1
Happy 670.15 (100.20) 628.38 (85.79) 647.83 (96.71) 611.60 (105.70) 756.14 (105.66) 768.56 (121.68) 780.77 (137.48) 734.14 (123.25) 681.73 (159.11) 618.07 (100.26) 623.77 (96.40) 608.18 (90.55)
Sad 639.75 (101.17) 585.02 (97.86) 595.80 (87.83) 576.68 (102.15) 725.06 (121.26) 706.72 (93.15) 720.50 (100.72) 691.43 (98.32) 608.03 (90.33) 574.65 (94.85) 585.00 (91.37) 579.84 (99.57)
Control 640.04 (106.70) 613.34 (93.09) 613.88 (96.07) 608.88 (100.73) 737.32 (109.15) 768.17 (125.21) 756.65 (133.40) 696.38 (99.11) 644.26 (117.93) 603.13 (92.20) 615.32 (101.39) 598.08 (79.44)
All 649.97 (101.80) 608.42 (92.65) 618.85 (94.38) 598.49 (102.35) 739.29 (111.19) 746.74 (115.51) 752.00 (124.74) 707.23 (107.32) 644.04 (126.78) 598.12 (95.94) 607.50 (96.08) 595.05 (89.77)
ANT-T2
Happy 641.62 (93.44) 565.55 (80.40) 600.29 (122.64) 575.50 (110.46) 742.90 (110.74) 707.47 (92.03) 727.43 (96.23) 678.41 (98.75) 619.69 (94.52) 565.44 (78.25) 583.76 (94.38) 575.37 (152.64)
Sad 608.10 (116.83) 565.31 (110.98) 549.73 (82.33) 559.23 (95.97) 690.73 (110.61) 680.12 (95.60) 678.49 (98.06) 651.91 (115.03) 617.50 (105.39) 543.95 (86.81) 561.45 (100.25) 535.99 (91.40)
Control 640.12 (129.14) 590.72 (119.17) 588.31 (105.63) 542.72 (77.42) 710.78 (113.59) 725.72 (128.31) 710.22 (118.55) 676.34 (128.69) 593.95 (78.15) 571.19 (114.36) 565.36 (88.21) 535.27 (71.32)
All 629.38 (112.92) 573.41 (103.48) 578.77 (104.93) 559.44 (95.10) 714.45 (111.76) 703.64 (105.74) 704.82 (104.57) 668.46 (113.15) 610.79 (92.89) 559.71 (92.94) 570.12 (93.45) 548.89 (110.57)
ANT: collapsed across time
Happy 655.89 (90.13) 596.96 (76.64) 624.06 (91.75) 593.55 (103.54) 749.52 (104.89) 738.02 (104.75) 754.10 (113.05) 706.27 (103.62) 650.71 (114.87) 591.76 (86.94) 603.76 (93.82) 591.77 (111.51)
Sad 623.92 (97.43) 575.16 (100.68) 572.77 (79.86) 567.96 (94.68) 707.90 (109.82) 693.42 (89.62) 699.49 (94.92) 671.67 (101.54) 612.77 (92.87) 559.30 (86.63) 573.22 (88.92) 557.91 (89.74)
Control 640.08 (114.34) 602.03 (92.65) 601.09 (92.92) 575.80 (72.67) 724.05 (97.66) 746.94 (114.11) 733.44 (115.67) 686.36 (102.27) 619.11 (87.07) 587.16 (95.31) 590.34 (85.51) 566.68 (69.04)
All 639.68 (99.89) 590.91 (89.85) 598.81 (89.17) 578.96 (90.60) 726.87 (104.06) 725.19 (103.79) 728.41 (108.43) 687.84 (101.65) 627.42 (98.75) 578.92 (89.16) 588.81 (89.14) 571.97 (91.59)
Table 4.

The full mixed ANOVA model results for reaction time across both ANT time points.

Effect, interaction, or contrast Statistic
Group F(2, 54) = 0.81, p = 0.45, ηp2 = 0.03
Time F(1, 54) = 23.71, p < 0.001*, ηp2 = 0.31
Time x group F(2, 54) = 0.41, p = 0.67, ηp2 = 0.02
Congruency F(1.60, 86.63) = 610.66, p < 0.001*, ηp2 = 0.91
Cue F(2.61, 140.78) = 57.45, p < 0.001*, ηp2 = 0.52
Congruency x group F(4, 108) = 0.73, p = 0.58, ηp2 = 0.03
Cue x group*** F(6, 162) = 2.33, p = 0.035*, ηp2 = 0.08
Time x congruency F(2, 108) = 0.57, p = 0.57, ηp2 = 0.10
Time x congruency x group F(4, 108) = 1.36, p = 0.25, ηp2 = 0.05
Time x cue F(3, 162) = 2.69, p = 0.048*, 0.05
Time x cue x group F(6, 162) = 1.01, p = 0.42, ηp2 = 0.04
Congruency x cue F(4.81, 259.87) = 8.02, p < 0.001*, ηp2 = 0.13
Congruency x cue x group F(12, 324) = 0.42, p = 0.96, ηp2 = 0.02
Time x congruency x cue F(4.44, 240.04) = 0.43, p = 0.81, ηp2 = 0.01
Time x congruency x cue x group F(12, 324) = 2.35, p = 0.007*, ηp2 = 0.08
Pairwise comparisons for the main effect of congruency
Executive control (incongruent vs. neutral) t(56) = 27.80, FDR p = 0.001**, d = 3.68
Incongruent vs. congruent t(56) = 27.66, FDR p = 0.001**, d = 3.66
Congruent vs. neutral t(56) = 4.25, FDR p = 0.001**, d = 0.56
Pairwise comparisons for the main effect of cue
Alerting (no cue vs. double cue) t(56) = 7.75, FDR p = 0.001**, d = 1.03
Orienting (center cue vs. valid spatial cue) t(56) = 5.90, FDR p = 0.001**, d = 0.78
No cue vs. center cue t(56) = 6.96, FDR p = 0.001**, d = 0.92
No cue vs. valid cue t(56) = 11.18, FDR p = 0.001**, d = 1.48
Double cue vs. center cue t(56) = 2.17, FDR p = 0.034**, d = 0.29
Double cue vs. valid cue t(56) = 4.61, FDR p = 0.001**, d = 0.61
Pairwise comparisons for the interaction: congruency x cue
Congruent: alerting t(56) = 7.31, FDR p = 0.002**, d = 0.97
Congruent: orienting t(56) = 3.06, FDR p = 0.005**, d = 0.41
Incongruent: alerting t(56) = 0.30, FDR p = 0.76, d = 0.04
Incongruent: orienting t(56) = 5.97, FDR p = 0.002**, d = 0.79
Neutral: alerting t(56) = 8.25, FDR p = 0.002**, d = 1.09
Neutral: orienting t(56) = 2.16, FDR p = 0.042**, d = 0.27

*Significant at p < 0.05; **Significant at FDR p < 0.05; ***All post-hoc comparisons were non-significant at p < 0.05. The alerting effect is characterized by faster responses on double cue compared to no cued trials. The orienting effect is characterized by faster responses on trials with valid spatial cues compared to trials with center cues.

The time x congruency x cue x group interaction was also significant [F(12, 273.93) = 2.35, p = 0.007]. As expected, no group demonstrated the expected alerting effect (faster responses on double cue than no cue trials) within incongruent trials at time one, and only the happy music listening group demonstrated the alerting effect within incongruent trials at time two. Within congruent and neutral trials, the alerting effect was present at time one and time two for both the happy and sad music groups. For the control group, the alerting effect was observed at time one for neutral trials, but not congruent trials. The alerting effect was not present at time two within congruent or neutral trials for the control group (Table 5; Figure 5). Supplementary Figure S1 depicts individual differences in how alerting changes within incongruent, congruent, and neutral trials from time one to time two for each participant.

Table 5.

Post-hoc comparisons within reaction time for the group x time x congruency x cue interaction.

Contrast Statistic
Happy music group
Time 1: incongruent: alerting t(18) = −1.55, FDR p = 0.19, d = 0.36
Time 2: incongruent: alerting t(18) = 2.76, FDR p = 0.022*, d = 0.63
Time 1: congruent: alerting t(18) = 3.52, FDR p = 0.006*, d = 0.81
Time 2: congruent: alerting t(18) = 6.11, FDR p = 0.006*, d = 1.40
Time 1: neutral: alerting t(18) = 3.52, FDR p = 0.006*, d = 0.81
Time 2: neutral: alerting t(18) = 3.77, FDR p = 0.006*, d = 0.87
Time 1: incongruent: orienting t(18) = 2.60, FDR p = 0.027*, d = 0.60
Time 2: incongruent: orienting t(18) = 3.33, FDR p = 0.008*, d = 0.76
Time 1: congruent: orienting t(18) = 3.42, FDR p = 0.007*, d = 0.78
Time 2: congruent: orienting t(18) = 0.86, FDR p = 0.44, d = 0.20
Time 1: neutral: orienting t(18) = 1.18, FDR p = 0.30, d = 0.27
Time 2: neutral: orienting t(18) = 0.25, FDR p = 0.80, d = 0.06
Sad music group
Time 1: incongruent: alerting t(19) = 1.49, FDR p = 0.20, d = 0.33
Time 2: incongruent: alerting t(19) = 1.22, FDR p = 0.29, d = 0.27
Time 1: congruent: alerting t(19) = 4.73, FDR p = 0.004*, d = 1.06
Time 2: congruent: alerting t(19) = 3.47, FDR p = 0.009*, d = 0.78
Time 1: neutral: alerting t(19) = 6.04, FDR p = 0.006*, d = 1.35
Time 2: neutral: alerting t(19) = 7.98, FDR p = 0.006*, d = 1.78
Time 1: incongruent: orienting t(19) = 3.09, FDR p = 0.01*, d = 0.69
Time 2: incongruent: orienting t(19) = 1.92, FDR p = 0.11, d = 0.43
Time 1: congruent: orienting t(19) = 2.69, FDR p = 0.03*, d = 0.60
Time 2: congruent: orienting t(19) = −0.86, FDR p = 0.44, d = 0.19
Time 1: neutral: orienting t(19) = 0.55, FDR p = 0.59, d = 0.12
Time 2: neutral: orienting t(19) = 2.08, FDR p = 0.09, d = 0.47
No music group
Time 1: incongruent: alerting** t(17) = −2.95, FDR p = 0.03*, d = 0.70
Time 2: incongruent: alerting t(17) = −0.99, FDR p = 0.37, d = 0.23
Time 1: congruent: alerting t(17) = 1.99, FDR p = 0.10, d = 0.47
Time 2: congruent: alerting t(17) = 2.30, FDR p = 0.06, d = 0.54
Time 1: neutral: alerting t(17) = 2.51, FDR p = 0.048*, d = 0.59
Time 2: neutral: alerting t(17) = 1.27, FDR p = 0.26, d = 0.30
Time 1: incongruent: orienting t(17) = 2.96, FDR p = 0.03*, d = 0.70
Time 2: incongruent: orienting t(17) = 1.51, FDR p = 0.20, d = 0.36
Time 1: congruent: orienting t(17) = 0.41, FDR p = 0.69, d = 0.10
Time 2: congruent: orienting t(17) = 3.47, FDR p = 0.02*, d = 0.82
Time 1: neutral: orienting t(17) = 1.87, FDR p = 0.12, d = 0.44
Time 2: neutral: orienting t(17) = 3.92, FDR p = 0.01*, d = 0.93

*Significant at FDR p < 0.05. **There is a significant difference between the no cue and double cue trials, but in the opposite direction of the alerting effect (i.e., no cue trials were faster than double cue trials). The alerting effect is faster responses on double cue compared to no cued trials. The orienting effect is faster responses on trials with valid spatial cues compared to trials with center cues.

Figure 5.

Figure 5

The alerting (no cue vs. double cue) effect separated by congruency (incongruent, congruent, and neutral) for the happy (A), sad (B), and no music (C) listening groups. Error bars represent ± 1 standard error. Note, within the control group, there is a significant difference between the no cue and double cue trials within neutral trials at time one. This difference is not marked as significant as it is in the opposite direction of the alerting effect (i.e., no cue trials were faster than double cue trials).

We also observed the orienting effect (faster responses on valid spatial cue than center cue trials) to interact with group, time, and congruency. Within incongruent trials, all groups demonstrated the orienting effect at time one, but this effect was only maintained at time two for the happy music listening group. Within congruent trials, the happy and sad music groups showed the orienting effect at time one, but not at time two. The control group did not demonstrate the orienting effect within congruent trials at time one but did at time two. The orienting effect was not observed within any group at time one within neutral trials, and only the control group demonstrated the orienting effect at time two (Table 5; Figure 6). Supplementary Figure S2 depicts individual differences in how orienting changes within incongruent, congruent, and neutral trials from time one to time two for each participant.

Figure 6.

Figure 6

The orienting (center cue vs. valid spatial cue) effect separated by congruency (incongruent, congruent, and neutral) for the happy (A), sad (B), and no music (C) listening groups. Error bars represent ± 1 standard error.

4. Discussion

It is well-established that attention, specifically alerting and executive control, decline with increasing age (Fernandez-Duque and Black, 2006; Jennings et al., 2007; Gamboz et al., 2010; Mahoney et al., 2010; Zhou et al., 2011; Knight and Mather, 2013; Kaufman et al., 2016; Williams et al., 2016), likely due to atrophy in the prefrontal cortex as part of the normal aging process (Peters, 2006; Glisky, 2007; Zanto and Gazzaley, 2019). However, few studies have investigated strategies to maintain attention in older adults. Listening to happy music may be a potential avenue as past research shows that music written in the major mode with a fast tempo increases attentional performance in older adults (Fernandez et al., 2019; Marti-Marca et al., 2020). The present study sought to investigate how listening to happy music (major mode, fast tempo), compared to sad music (minor mode, slow tempo) or no music, acutely impacts alerting, orienting, and executive control attention in older adults. To this end, alerting and executive control improved after listening to happy and sad music, respectively. These results are discussed in their respective sections below.

4.1. Alerting and executive control interact before music listening

Consistent with previous literature (Jennings et al., 2007; Gamboz et al., 2010; Zhou et al., 2011), the mixed ANOVA revealed that all participants demonstrated the expected alerting (i.e., faster reaction times on double cue versus no cue trials), orienting (i.e., faster reaction times on valid spatial cue compared to center cued trials), and executive control effects (i.e., slower reaction times on incongruent trials than neutral trials) collapsed across time. The congruency x cue interaction was also significant across time: the alerting effect was present when the flanker task was congruent or neutral, but not when it was incongruent. This finding aligns with previous work and suggests that the attention networks are not entirely independent (Callejas et al., 2004; Fan et al., 2009; Gamboz et al., 2010; Ishigami and Klein, 2010; McConnell and Shore, 2011a; Weinbach and Henik, 2012).

This diminished alerting effect within incongruent trials may be attributed to alerting and executive control utilizing partially shared resources (Fan et al., 2005; Roberts and Hall, 2008; Anderson et al., 2010; Petersen and Posner, 2012; Rinne et al., 2013), as well as the overall limited capacity of attention (Kahneman, 1973). A warning cue plus an incongruent trial may cause older adults to exceed their resource capacity as the warning cue uses more attention resources than no cue. This may subsequently result in reduced performance, or longer response times on the trials in which a warning cue precedes an incongruent trial. The speed-accuracy tradeoff may also contribute to the diminished alerting effect within incongruent trials. Older adults are known to favor accuracy over speed, and this effect is exacerbated on harder tasks (Heitz, 2014). The speed-accuracy tradeoff may specifically affect incongruent trials in that no matter the type of cueing provided, older adults take the necessary time to accurately complete the task, resulting in similar response times across all incongruent trials.

4.2. Music interacts with attention

4.2.1. Alerting

Only one study to our knowledge has investigated the impact of music listening prior to completing the ANT. Their results showed that the alerting effect was similar for younger adults who listened to happy (major mode, fast tempo) and sad music (minor mode, slow tempo; McConnell and Shore, 2011b). Our results expand upon this work to show that in older adults, alerting attention increases in response to happy music listening, but not after listening to sad music or no music. Furthermore, when we focus on incongruent trials, the trials most challenging for older adults (Mutter et al., 2005; Ward et al., 2021), we show that the alerting effect was not significant for the happy music group at time one, but was at time two. This same pattern within incongruent trials was not observed for the sad and no music listening groups. These findings correspond to the arousal-mood hypothesis, which posits that listening to non-lyrical music written in the major mode with a fast tempo, induces positive mood and increases levels of arousal more so than sad music or music written in the minor mode with a slow tempo (Balch et al., 1999; Gabrielsson and Lindström, 2001; Thompson et al., 2001; Husain et al., 2002; McConnell and Shore, 2011b). These increases in mood and arousal following happy music listening subsequently result in superior task performance (Steele, 2000; Thompson et al., 2001; Husain et al., 2002; Schellenberg and Hallam, 2005; Schellenberg et al., 2007; Schellenberg, 2012).

In addition to increasing mood and arousal levels, happy music may also improve alerting attention by stimulating its supporting neural resources. Fernandez et al. (2019) found that older adults who listened to happy background music had better alerting attention than those who listened to sad music. They further observed increased activation in the prefrontal cortex for happy music only. Increased activation in the prefrontal cortex during happy music listening was attributed to executive control (Fernandez et al., 2019). However, it could also be associated with alerting attention as fMRI has a relatively poor temporal resolution (Glover, 2011; Loued-Khenissi et al., 2019; Puttaert et al., 2020), which may make it difficult to parse the alerting subsystem from the executive control subsystem given the short interstimulus period between the cue offset and the target onset (i.e., 400 milliseconds). Regardless, music listening activates areas of the brain associated with attention (Mitterschiffthaler et al., 2007; Brattico et al., 2011). Music listening also promotes neural entrainment, specifically in response to musical features such as rhythm and pitch (Nozaradan et al., 2011, 2012; Doelling and Poeppel, 2015; Wollman et al., 2020). Together these findings suggest a potential therapeutic benefit to happy music listening in the short-term, and further show that in some instances happy music may be able to reinstate a previously absent or diminished alert state in older adults. However, additional work is needed to confirm and extend these findings using a more diverse group of participants.

4.2.2. Orienting

The paired samples t-tests indicate that the orienting effect did not change in response to music listening. However, further investigation of the orienting effect within the challenging incongruent trials indicates that the orienting effect was significant at time one for all groups, however, was only significant in the happy music listening group at time two. This finding, which indicates that happy music may be modulating orienting attention contrasts with the previous literature showing that orienting attention does not change in response to happy or sad music (Jiang et al., 2011; McConnell and Shore, 2011b; Fernandez et al., 2019). Yet, the majority of research investigating the impact of music listening prior to task completion has utilized visuospatial tasks (Rauscher et al., 1993, 1995; Thompson et al., 2001; Husain et al., 2002; Schellenberg and Hallam, 2005; Schellenberg et al., 2007; Schellenberg, 2012), and the orienting cue provides relevant information about the visuospatial location of the flanker task. Music listening prior to ANT performance may therefore increase participants’ ability to effectively respond to both valid spatial and centrally cued trials, possibly because of increased alertness (Fan et al., 2002). However, future work is needed to explore this hypothesis as alerting and orienting did not correlate (r(57) = 0.03, p = 0.81), nor interact in this study or others (Fan et al., 2002; Yin et al., 2012; Spagna et al., 2015; Aminihajibashi et al., 2020).

4.2.3. Executive control

The arousal-mood hypothesis suggests that executive control performance should improve after listening to happy music (Thompson et al., 2001; Husain et al., 2002). Older adults have also been shown to perform better on the executive control aspect of the ANT when they listened to happy music compared to sad or no music (Fernandez et al., 2019). Therefore, our finding that executive control attention improved after listening to sad music was unexpected, but does align with work showing that sad music, or music written in the minor mode, can increase executive control attention more so than happy music, or music written in the major mode, particularly when arousal levels are high (McConnell and Shore, 2011b). Listening to sad music may increase executive control attention more so than listening to happy music because sad music focuses attention (Jiang et al., 2011) while happy music broadens attention (Rowe et al., 2007). This broadening of attention subsequently impairs inhibitory control (Rowe et al., 2007). Thus, sad music may increase executive control by narrowing the focus of attention, such that participants are more locally attentive to the information they are cued to attend to rather than globally attentive to all aspects of the flanker task, including the distractor information within the incongruent trials (Rowe et al., 2007; Jiang et al., 2011; McConnell and Shore, 2011b). Sad music may also prime participants to internally focus their attention (Jiang et al., 2011; Taruffi et al., 2017), resulting in improved conflict resolution (Lippelt et al., 2014). However, future work is needed to confirm this relationship between sad music listening and executive control performance as group did not interact with congruency or time in the mixed ANOVA.

4.3. Musical considerations and future directions

Overall, this study demonstrates that listening to music increases attentional performance, particularly alerting and executive control. This conclusion may only apply to females, who comprised most of our sample. The majority of research indicates that attention does not differ between females and males (Mahoney et al., 2010; Zhou et al., 2011; McDonough et al., 2019). However, males and females perceive emotion differently (Aljanaki et al., 2016; Cloutier et al., 2020), with females reporting higher levels of happiness than males (Cloutier et al., 2020). It is therefore possible that our predominately female sample contributed to the sad music listening group rating their musical composition more neutrally (44.8/100) than what would be expected based on the arousal-mood hypothesis (Husain et al., 2002), and the work done by Trost et al. (2012). Age may also be influencing participant’s emotional perception of each composition. Music is perceived more positively with increasing age, regardless of mode, key, and tempo (Cohrdes et al., 2020). This may result in older adults finding a wider array of music, including those composed in the minor mode (like our sad composition), to be happier than what is perceived by younger adults.

The sad music could also have been rated more neutrally by our older adult participants because of differences in exposure to the composition style of each musical composition. For instance, our non-musician participants likely have more experience with the happy music’s composition style as its style, 19th Century Romanticism, is one we are accustomed to listening to in movies and commercials in the United States. Similarly, participants’ decreased experience with the sad music composition’s style may have caused them to perceive the sad music as more surprising, resulting in a more neutral emotional rating. These two explanations could explain why we did not observe more robust changes in attention between the happy and sad music listening groups. Future research, with a more balanced participant sample, is needed to better understand how middle and older aged adults perceive emotion in music, and how their perceptions may be modulated by gender and different musical features. A more balanced participant sample will also allow future studies to address whether the observed changes in attention following music listening differ when stratified by variables such as education and gender (see Supplementary material for exploratory analyses looking at the influence of education, age, and gender on the relationships between happy music and alerting and sad music and executive control in the current participant sample). It would also be useful for research to explore whether music listening similarly impacts attention in clinical populations which show impairments in alerting, orienting, and executive control (e.g., stroke).

While we find that happy and sad music both improve different aspects of attention, additional research is needed to investigate what role tempo and composition mode play in improving attention in older adults as we did not vary these features beyond the specified pairings (major-fast, minor-slow). We therefore cannot distinguish whether composition mode (major vs. minor), tempo (fast vs. slow), or the combination was inducing the observed changes in alerting and executive control. Similarly, other aspects of music not controlled for in this study including timbre, tonality, harmony, and composition key (to name a few) could also be contributing to the observed changes in attention. While a complete discussion of these aspects of music and their relationship to attention is beyond the scope of this study, it is important to note that our results may not generalize to all music written in the major mode with a fast tempo or minor mode with a slow tempo as it is possible that other musical properties are driving changes in attention. However, we speculate that mode and tempo may be important contributors to the observed changes in attention as the major mode and fast tempo and minor mode and slow tempo pairings are readily associated with feelings of happiness and sadness, and high and low arousal in non-musicians (Husain et al., 2002; Webster and Weir, 2005; Gabrielsson and Lindström, 2010; Hunter et al., 2010; Trost et al., 2012). Nonetheless, future work is needed to systematically study how aspects of music, beyond mode and tempo, modulate older adults’ attention.

4.4. Conclusion

Our results suggest that listening to both happy and sad music impacts middle-aged and older adults’ attention. More specifically, we found that happy music increases alerting attention, particularly on trials which contain conflict. Orienting attention also appeared to be modulated by happy music within incongruent trials only. These findings align with the arousal-mood hypothesis and provide further evidence that happy music, written in the major mode with a fast tempo, increases cognitive performance by increasing arousal (or alertness). Unexpectedly, sad music, written in the minor mode with a slow tempo, appears to improve executive control performance, possibly by priming the participant to focus their attention internally, thus aiding conflict resolution. While future studies are needed to replicate our findings and further explore the relationship between attention and other aspects of music beyond mode and tempo, our findings indicate that music written in the major mode with a fast tempo (happy) and minor mode with a slow tempo (sad) acutely modulate different aspects of attention.

Data availability statement

The datasets presented in this study can be found in online repositories. The names of the repository/repositories and accession number(s) can be found at: https://osf.io/gf3km/.

Ethics statement

The studies involving human participants were reviewed and approved by Midwestern University’s Institutional Review Board. The patients/participants provided their written informed consent to participate in this study.

Author contributions

ND and AL was involved in experimental design, participant recruitment and testing, data analysis, and manuscript writing. SB, NJ, and IR were involved in experimental design, data analysis, and manuscript editing. All authors contributed to the article and approved the submitted version.

Funding

This research was supported by Midwestern University.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary material

The Supplementary material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fpsyg.2023.1029773/full#supplementary-material

Footnotes

1The invalid cues were not included in the analysis as they are not used in the calculation of any of the attentional subsystems. However, the means and standard deviations are included for completeness (Table 2).

References

  1. Aljanaki A., Wiering F., Veltkamp R. C. (2016). Studying emotion induced by music through a crowdsourcing game. Inf. Process. Manag. 52, 115–128. doi: 10.1016/j.ipm.2015.03.004 [DOI] [Google Scholar]
  2. Aminihajibashi S., Hagen T., Laeng B., Espeseth T. (2020). Pupillary and behavioral markers of alerting and orienting: an individual difference approach. Brain Cogn. 143:105597. doi: 10.1016/j.bandc.2020.105597, PMID: [DOI] [PubMed] [Google Scholar]
  3. Anderson J. S., Ferguson M. A., Lopez-Larson M., Yurgelun-Todd D. (2010). Topographic maps of multisensory attention. Proc. Natl. Acad. Sci. U. S. A. 107, 20110–20114. doi: 10.1073/pnas.1011616107, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
  4. Balch W. R., Myers D. M., Papotto C. (1999). Dimensions of mood in mood-dependent memory. J. Exp. Psychol. Learn. Mem. Cogn. 25, 70–83. [DOI] [PubMed] [Google Scholar]
  5. Benjamini Y., Hochberg Y. (1995). Controlling the false discovery rate: a practical and powerful approach to multiple testing. J. R. Stat. Soc. Ser. B 57, 289–300. [Google Scholar]
  6. Brattico E., Alluri V., Bogert B., Jacobsen T., Vartiainen N., Nieminen S., et al. (2011). A functional MRI study of happy and sad emotions in music with and without lyrics. Front. Psychol. 2:308. doi: 10.3389/fpsyg.2011.00308 [DOI] [PMC free article] [PubMed] [Google Scholar]
  7. Bruyneel L., van Steenbergen H., Hommel B., Band G. P., De Raedt R., Koster E. H. (2013). Happy but still focused: failures to find evidence for a mood-induced widening of visual attention. Psychol. Res. 77, 320–332. doi: 10.1007/s00426-012-0432-1, PMID: [DOI] [PubMed] [Google Scholar]
  8. Callejas A., Lupiáñez J., Tudela P. (2004). The three attentional networks: on their independence and interactions. Brain Cogn. 54, 225–227. doi: 10.1016/j.bandc.2004.02.012, PMID: [DOI] [PubMed] [Google Scholar]
  9. Cassidy G., MacDonald R. A. R. (2007). The effect of background music and background noise on the task performance of introverts and extraverts. Psychol. Music 35, 517–537. doi: 10.1177/0305735607076444 [DOI] [Google Scholar]
  10. Chew A. S.-Q., Yu Y.-T., Chua S.-W., Gan S. K.-E. (2016). The effects of familiarity and language of background music on working memory and language tasks in Singapore. Psychol. Music 44, 1431–1438. doi: 10.1177/0305735616636209 [DOI] [Google Scholar]
  11. Cloutier A., Fernandez N. B., Houde-Archambault C., Gosselin N. (2020). Effect of background music on attentional control in older and young adults. Front. Psychol. 11:557225. doi: 10.3389/fpsyg.2020.557225, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
  12. Cohrdes C., Wrzus C., Wald-Fuhrmann M., Riediger M. (2020). “The sound of affect”: age differences in perceiving valence and arousal in music and their relation to music characteristics and momentary mood. Music. Sci. 24, 21–43. doi: 10.1177/1029864918765613 [DOI] [Google Scholar]
  13. Corbetta M., Shulman G. L. (2002). Control of goal-directed and stimulus-driven attention in the brain. Nat. Rev. Neurosci. 3, 201–215. doi: 10.1038/nrn755 [DOI] [PubMed] [Google Scholar]
  14. Doelling K. B., Poeppel D. (2015). Cortical entrainment to music and its modulation by expertise. Proc. Natl. Acad. Sci. 112, E6233–E6242. doi: 10.1073/pnas.1508431112, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
  15. Fan J., Gu X., Guise K. G., Liu X., Fossella J., Wang H., et al. (2009). Testing the behavioral interaction and integration of attentional networks. Brain Cogn. 70, 209–220. doi: 10.1016/j.bandc.2009.02.002, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
  16. Fan J., McCandliss B. D., Fossella J., Flombaum J. I., Posner M. I. (2005). The activation of attentional networks. NeuroImage 26, 471–479. doi: 10.1016/j.neuroimage.2005.02.004 [DOI] [PubMed] [Google Scholar]
  17. Fan J., McCandliss B. D., Sommer T., Raz A., Posner M. I. (2002). Testing the efficiency and independence of attentional networks. J. Cogn. Neurosci. 14, 340–347. doi: 10.1162/089892902317361886, PMID: [DOI] [PubMed] [Google Scholar]
  18. Fernandez N. B., Trost W. J., Vuilleumier P. (2019). Brain networks mediating the influence of background music on selective attention. Soc. Cogn. Affect. Neurosci. 14, 1441–1452. doi: 10.1093/scan/nsaa004, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
  19. Fernandez-Duque D., Black S. E. (2006). Attentional networks in normal aging and Alzheimer's disease. Neuropsychology 20, 133–143. doi: 10.1037/0894-4105.20.2.133 [DOI] [PubMed] [Google Scholar]
  20. Finucane A. M., Whiteman M. C., Power M. J. (2010). The effect of happiness and sadness on alerting, orienting, and executive attention. J. Atten. Disord. 13, 629–639. doi: 10.1177/1087054709334514, PMID: [DOI] [PubMed] [Google Scholar]
  21. Gabrielsson A., Lindström E. (2001). “The influence of musical structure on emotional expression” in Music and Emotion: Theory and Research. eds. Juslin P. N., Sloboda J. A. (Oxford, United Kingdom: Oxford University Press; ), 223–248. [Google Scholar]
  22. Gabrielsson A., Lindström E. (2010). “The role of structure in the musical expression of emotions” in Handbook of Music and Emotion: Theory, Research, Applications. eds. Juslin P. N., Sloboda J. A. (Oxford, United Kingdom: Oxford University Press; ), 367–400. [Google Scholar]
  23. Gamboz N., Zamarian S., Cavallero C. (2010). Age-related differences in the attention network test (ANT). Exp. Aging Res. 36, 287–305. doi: 10.1080/0361073X.2010.484729, PMID: [DOI] [PubMed] [Google Scholar]
  24. Glisky E. L. (2007). “Changes in cognitive function in human aging” in Brain aging: Models, Methods, and Mechanisms. ed. Riddle D. M. (Boca Raton, FL: CRC Press, Routledge, Taylor & Francis Group; ), 3–20. [Google Scholar]
  25. Glover G. H. (2011). Overview of functional magnetic resonance imaging. Neurosurg. Clin. N. Am. 22, 133–139. doi: 10.1016/j.nec.2010.11.001, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
  26. Gonzalez M. F., Aiello J. R. (2019). More than meets the ear: investigating how music affects cognitive task performance. J. Exp. Psychol. Appl. 25, 431–444. doi: 10.1037/xap0000202, PMID: [DOI] [PubMed] [Google Scholar]
  27. Hart S. G., Staveland L. E. (1988). “Development of NASA-TLX (task load index): results of empirical and theoretical research” in Human Mental Workload. eds. Hancock P. A., Meshkati N. (Amsterdam, Netherlands: Elsevier; ), 139–183. [Google Scholar]
  28. Harvey P. D. (2019). Domains of cognition and their assessment. Dialogues Clin. Neurosci. 21, 227–237. doi: 10.31887/DCNS.2019.21.3/pharvey, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
  29. Heitz R. P. (2014). The speed-accuracy tradeoff: history, physiology, methodology, and behavior. Front. Neurosci. 8:150. doi: 10.3389/fnins.2014.00150 [DOI] [PMC free article] [PubMed] [Google Scholar]
  30. Hoe J., Hancock G., Livingston G., Woods B., Challis D., Orrell M. (2009). Changes in the quality of life of people with dementia living in care homes. Alzheimer Dis. Assoc. Disord. 23, 285–290. doi: 10.1097/WAD.0b013e318194fc1e, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
  31. Hunter P. G., Schellenberg E. G., Schimmack U. (2010). Feelings and perceptions of happiness and sadness induced by music: similarities, differences, and mixed emotions. Psychol. Aesthet. Creat. Arts 4, 47–56. doi: 10.1037/a0016873 [DOI] [Google Scholar]
  32. Husain G., Thompson W. F., Schellenberg E. G. (2002). Effects of musical tempo and mode on arousal, mood, and spatial abilities. Music. Percept. 20, 151–171. doi: 10.1525/mp.2002.20.2.151 [DOI] [Google Scholar]
  33. IBM (2020). IBM SPSS Statistics for Windows, Version 27.0. Armonk, NY: IBM Corp. [Google Scholar]
  34. Inzlicht M., Shenhav A., Olivola C. Y. (2018). The effort paradox: effort is both costly and valued. Trends Cogn. Sci. 22, 337–349. doi: 10.1016/j.tics.2018.01.007, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
  35. Ishigami Y., Klein R. M. (2010). Repeated measurement of the components of attention using two versions of the attention network test (ANT): stability, isolability, robustness, and reliability. J. Neurosci. Methods 190, 117–128. doi: 10.1016/j.jneumeth.2010.04.019, PMID: [DOI] [PubMed] [Google Scholar]
  36. Jennings J. M., Dagenbach D., Engle C. M., Funke L. J. (2007). Age-related changes and the attention network task: an examination of alerting, orienting, and executive function. Neuropsychology, development, and cognition. Section B. Aging Neuropsychol. Cognit. 14, 353–369. doi: 10.1080/13825580600788837, PMID: [DOI] [PubMed] [Google Scholar]
  37. Jernigan T. L., Archibald S. L., Fennema-Notestine C., Gamst A. C., Stout J. C., Bonner J., et al. (2001). Effects of age on tissues and regions of the cerebrum and cerebellum. Neurobiol. Aging 22, 581–594. doi: 10.1016/S0197-4580(01)00217-2 [DOI] [PubMed] [Google Scholar]
  38. Jiang J., Scolaro A. J., Bailey K., Chen A. (2011). The effect of music-induced mood on attentional networks. Int. J. Psychol. 46, 214–222. doi: 10.1080/00207594.2010.541255, PMID: [DOI] [PubMed] [Google Scholar]
  39. Kahneman D. (1973). Attention and Effort. London: Inform Healthcare. [Google Scholar]
  40. Kaufman D. A., Sozda C. N., Dotson V. M., Perlstein W. M. (2016). An event-related potential investigation of the effects of age on alerting, orienting, and executive function. Front. Aging Neurosci. 8:99. doi: 10.3389/fnagi.2016.00099, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
  41. Kiss L., Linnell K. J. (2021). The effect of preferred background music on task-focus in sustained attention. Psychol. Res. 85, 2313–2325. doi: 10.1007/s00426-020-01400-6, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
  42. Knight M., Mather M. (2013). Look out-it's your off-peak time of day! Time of day matters more for alerting than for orienting or executive attention. Exp. Aging Res. 39, 305–321. doi: 10.1080/0361073X.2013.779197, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
  43. Koelsch S. (2010). Towards a neural basis of music-evoked emotions. Trends Cogn. Sci. 14, 131–137. doi: 10.1016/j.tics.2010.01.002, PMID: [DOI] [PubMed] [Google Scholar]
  44. Koelsch S. (2014). Brain correlates of music-evoked emotions. Nat. Rev. Neurosci. 15, 170–180. doi: 10.1038/nrn3666, PMID: [DOI] [PubMed] [Google Scholar]
  45. Lippelt D. P., Hommel B., Colzato L. S. (2014). Focused attention, open monitoring and loving kindness meditation: effects on attention, conflict monitoring, and creativity - a review. Front. Psychol. 5:1083. doi: 10.3389/fpsyg.2014.01083, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
  46. Long X., Liao W., Jiang C., Liang D., Qiu B., Zhang L. (2012). Healthy aging: an automatic analysis of global and regional morphological alterations of human brain. Acad. Radiol. 19, 785–793. doi: 10.1016/j.acra.2012.03.006 [DOI] [PubMed] [Google Scholar]
  47. Loued-Khenissi L., Döll O., Preuschoff K. (2019). An overview of functional magnetic resonance imaging techniques for organizational research. Organ. Res. Methods 22, 17–45. doi: 10.1177/1094428118802631 [DOI] [Google Scholar]
  48. Lu H., Chan S. S., Fung A. W., Lam L. C. (2016). Efficiency of attentional components in elderly with mild neurocognitive disorders shown by the attention network test. Dement. Geriatr. Cogn. Disord. 41, 93–98. doi: 10.1159/000441350, PMID: [DOI] [PubMed] [Google Scholar]
  49. Mahoney J. R., Verghese J., Goldin Y., Lipton R., Holtzer R. (2010). Alerting, orienting, and executive attention in older adults. J. Int. Neuropsychol. Soc. 16, 877–889. doi: 10.1017/S1355617710000767, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
  50. Mammarella N., Fairfield B., Cornoldi C. (2007). Does music enhance cognitive performance in healthy older adults? The Vivaldi effect. Aging Clin. Exp. Res. 19, 394–399. doi: 10.1007/BF03324720, PMID: [DOI] [PubMed] [Google Scholar]
  51. Marti-Marca A., Nguyen T., Grahn J. A. (2020). Keep calm and pump up the jams: how musical mood and arousal affect visual attention. Music Sci. 3:205920432092273. doi: 10.1177/2059204320922737 [DOI] [Google Scholar]
  52. McCabe D. P., Roediger H. L., McDaniel M. A., Balota D. A., Hambrick D. Z. (2010). The relationship between working memory capacity and executive functioning: evidence for a common executive attention construct. Neuropsychology 24, 222–243. doi: 10.1037/a0017619, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
  53. McConnell M. M., Shore D. I. (2011a). Mixing measures: testing an assumption of the attention network test. Atten. Percept. Psychophys. 73, 1096–1107. doi: 10.3758/s13414-010-0085-3, PMID: [DOI] [PubMed] [Google Scholar]
  54. McConnell M. M., Shore D. I. (2011b). Upbeat and happy: arousal as an important factor in studying attention. Cognit. Emot. 25, 1184–1195. doi: 10.1080/02699931.2010.524396, PMID: [DOI] [PubMed] [Google Scholar]
  55. McCutcheon L. E. (2000). Another failure to generalize the Mozart effect. Psychol. Rep. 87, 325–330. doi: 10.2466/pr0.2000.87.1.325, PMID: [DOI] [PubMed] [Google Scholar]
  56. McDonough I. M., Wood M. M., Miller W. S. (2019). A review on the trajectory of attentional mechanisms in aging and the Alzheimer's disease continuum through the attention network test. Yale J. Biol. Med. 92, 37–51. [PMC free article] [PubMed] [Google Scholar]
  57. Medina D., Barraza P. (2019). Efficiency of attentional networks in musicians and non-musicians. Heliyon 5:e01315. doi: 10.1016/j.heliyon.2019.e01315, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
  58. Mirsky A. F., Anthony B. J., Duncan C. C., Ahearn M. B., Kellam S. G. (1991). Analysis of the elements of attention: a neuropsychological approach. Neuropsychol. Rev. 2, 109–145. doi: 10.1007/BF01109051 [DOI] [PubMed] [Google Scholar]
  59. Mitterschiffthaler M. T., Fu C. H., Dalton J. A., Andrew C. M., Williams S. C. (2007). A functional MRI study of happy and sad affective states induced by classical music. Hum. Brain Mapp. 28, 1150–1162. doi: 10.1002/hbm.20337, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
  60. Mutter S. A., Naylor J. C., Patterson E. R. (2005). The effects of age and task context on Stroop task performance. Mem. Cogn. 33, 514–530. doi: 10.3758/BF03193068, PMID: [DOI] [PubMed] [Google Scholar]
  61. Nasreddine Z. S., Phillips N. A., Bédirian V., Charbonneau S., Whitehead V., Collin I., et al. (2005). The Montreal cognitive assessment, MoCA: a brief screening tool for mild cognitive impairment. J. Am. Geriatr. Soc. 53, 695–699. doi: 10.1111/j.1532-5415.2005.53221.x, PMID: [DOI] [PubMed] [Google Scholar]
  62. Nozaradan S., Peretz I., Missal M., Mouraux A. (2011). Tagging the neuronal entrainment to beat and meter. J. Neurosci. 31, 10234–10240. doi: 10.1523/JNEUROSCI.0411-11.2011, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
  63. Nozaradan S., Peretz I., Mouraux A. (2012). Selective neuronal entrainment to the beat and meter embedded in a musical rhythm. J. Neurosci. 32, 17572–17581. doi: 10.1523/JNEUROSCI.3203-12.2012, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
  64. Palmiero M., Nori R., Rogolino C., D'Amico S., Piccardi L. (2016). Sex differences in visuospatial and navigational working memory: the role of mood induced by background music. Exp. Brain Res. 234, 2381–2389. doi: 10.1007/s00221-016-4643-3, PMID: [DOI] [PubMed] [Google Scholar]
  65. Peters R. (2006). Ageing and the brain. Postgrad. Med. J. 82, 84–88. doi: 10.1136/pgmj.2005.036665, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
  66. Petersen S. E., Posner M. I. (2012). The attention system of the human brain: 20 years after. Annu. Rev. Neurosci. 35, 73–89. doi: 10.1146/annurev-neuro-062111-150525, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
  67. Posner M. I., Petersen S. E. (1990). The attention system of the human brain. Annu. Rev. Neurosci. 13, 25–42. doi: 10.1146/annurev.ne.13.030190.000325 [DOI] [PubMed] [Google Scholar]
  68. Puttaert D., Coquelet N., Wens V., Peigneux P., Fery P., Rovai A., et al. (2020). Alterations in resting-state network dynamics along the Alzheimer's disease continuum. Sci. Rep. 10:21990. doi: 10.1038/s41598-020-76201-3, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
  69. Ransdell S. E., Gilroy L. (2001). The effects of background music on word processed writing. Comput. Hum. Behav. 17, 141–148. doi: 10.1016/S0747-5632(00)00043-1 [DOI] [Google Scholar]
  70. Rauscher F. H., Shaw G. L., Ky K. N. (1993). Music and spatial task performance. Nature 365:611. doi: 10.1038/365611a0 [DOI] [PubMed] [Google Scholar]
  71. Rauscher F. H., Shaw G. L., Ky K. N. (1995). Listening to Mozart enhances spatial-temporal reasoning: towards a neurophysiological basis. Neurosci. Lett. 185, 44–47. doi: 10.1016/0304-3940(94)11221-4, PMID: [DOI] [PubMed] [Google Scholar]
  72. Rey G. D. (2012). A review of research and a meta-analysis of the seductive detail effect. Educ. Res. Rev. 7, 216–237. doi: 10.1016/j.edurev.2012.05.003 [DOI] [Google Scholar]
  73. Rinne P., Hassan M., Goniotakis D., Chohan K., Sharma P., Langdon D., et al. (2013). Triple dissociation of attention networks in stroke according to lesion location. Neurology 81, 812–820. doi: 10.1212/WNL.0b013e3182a2ca34, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
  74. Roberts K. L., Hall D. A. (2008). Examining a supramodal network for conflict processing: a systematic review and novel functional magnetic resonance imaging data for related visual and auditory stroop tasks. J. Cogn. Neurosci. 20, 1063–1078. doi: 10.1162/jocn.2008.20074, PMID: [DOI] [PubMed] [Google Scholar]
  75. Román-Caballero R., Arnedo M., Triviño M., Lupiáñez J. (2018). Musical practice as an enhancer of cognitive function in healthy aging - a systematic review and meta-analysis. PLoS One 13:e0207957. doi: 10.1371/journal.pone.0207957, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
  76. Rowe G., Hirsh J. B., Anderson A. K. (2007). Positive affect increases the breadth of attentional selection. Proc. Natl. Acad. Sci. U. S. A. 104, 383–388. doi: 10.1073/pnas.0605198104, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
  77. Saint-Mont U. (2015). Randomization does not help much, comparability does. PLoS One 10:e0132102. doi: 10.1371/journal.pone.0132102, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
  78. Scahill R. I., Frost C., Jenkins R., Whitwell J. L., Rossor M. N., Fox N. C. (2003). A longitudinal study of brain volume changes in normal aging using serial registered magnetic resonance imaging. Arch. Neurol. 60, 989–994. doi: 10.1001/archneur.60.7.989, PMID: [DOI] [PubMed] [Google Scholar]
  79. Schellenberg E. G. (2012). “Cognitive performance after listening to music: a review of the Mozart effect” in Music, Health, and Wellbeing. eds. MacDonald R. A. R., Kreutz G., Mitchell L. (Oxford, United Kingdom: Oxford University Press; ), 325–338. [Google Scholar]
  80. Schellenberg E. G., Hallam S. (2005). Music listening and cognitive abilities in 10-and 11-year-olds: the blur effect. Ann. N. Y. Acad. Sci. 1060, 202–209. doi: 10.1196/annals.1360.013, PMID: [DOI] [PubMed] [Google Scholar]
  81. Schellenberg E. G., Nakata T., Hunter P. G., Tamoto S. (2007). Exposure to music and cognitive performance: tests of children and adults. Psychol. Music 35, 5–19. doi: 10.1177/0305735607068885 [DOI] [Google Scholar]
  82. Sohlberg M. M., Mateer C. A. (1987). Effectiveness of an attention-training program. J. Clin. Exp. Neuropsychol. 9, 117–130. doi: 10.1080/01688638708405352 [DOI] [PubMed] [Google Scholar]
  83. Spagna A., Mackie M. A., Fan J. (2015). Supramodal executive control of attention. Front. Psychol. 6:65. doi: 10.3389/fpsyg.2015.00065 [DOI] [PMC free article] [PubMed] [Google Scholar]
  84. Steele K. M. (2000). Arousal and mood factors in the "Mozart effect". Percept. Mot. Skills 91, 188–190. doi: 10.2466/pms.2000.91.1.188, PMID: [DOI] [PubMed] [Google Scholar]
  85. Steele K. M., Ball T. N., Runk R. (1997). Listening to Mozart does not enhance backwards digit span performance. Percept. Mot. Skills 84, 1179–1184. doi: 10.2466/pms.1997.84.3c.1179, PMID: [DOI] [PubMed] [Google Scholar]
  86. Steele K. M., Brown J. D., Stoecker J. A. (1999). Failure to confirm the Rauscher and Shaw description of recovery of the Mozart effect. Percept. Mot. Skills 88, 843–848. doi: 10.2466/pms.1999.88.3.843, PMID: [DOI] [PubMed] [Google Scholar]
  87. Svennerholm L., Boström K., Jungbjer B. (1997). Changes in weight and compositions of major membrane components of human brain during the span of adult human life of swedes. Acta Neuropathol. 94, 345–352. doi: 10.1007/s004010050717, PMID: [DOI] [PubMed] [Google Scholar]
  88. Taruffi L., Pehrs C., Skouras S., Koelsch S. (2017). Effects of sad and happy music on mind-wandering and the default mode network. Sci. Rep. 7:14396. doi: 10.1038/s41598-017-14849-0, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
  89. Thompson W. F., Schellenberg E. G., Husain G. (2001). Arousal, mood, and the Mozart effect. Psychol. Sci. 12, 248–251. doi: 10.1111/1467-9280.00345, PMID: [DOI] [PubMed] [Google Scholar]
  90. Trollor J. N., Valenzuela M. J. (2001). Brain ageing in the new millennium. Aust. N. Z. J. Psychiatry 35, 788–805. doi: 10.1046/j.1440-1614.2001.00969.x [DOI] [PubMed] [Google Scholar]
  91. Trost W., Ethofer T., Zentner M., Vuilleumier P. (2012). Mapping aesthetic musical emotions in the brain. Cereb. Cortex 22, 2769–2783. doi: 10.1093/cercor/bhr353, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
  92. Walhovd K. B., Fjell A. M., Reinvang I., Lundervold A., Dale A. M., Eilertsen D. E., et al. (2005). Effects of age on volumes of cortex, white matter and subcortical structures. Neurobiol. Aging 26, 1261–1270. doi: 10.1016/j.neurobiolaging.2005.05.020 [DOI] [PubMed] [Google Scholar]
  93. Ward N., Hussey E., Alzahabi R., Gaspar J. G., Kramer A. F. (2021). Age-related effects on a novel dual-task Stroop paradigm. PLoS One 16:e0247923. doi: 10.1371/journal.pone.0247923, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
  94. Webster G. D., Weir C. G. (2005). Emotional responses to music: interactive effects of mode, texture, and tempo. Motiv. Emot. 29, 19–39. doi: 10.1007/s11031-005-4414-0 [DOI] [Google Scholar]
  95. Weinbach N., Henik A. (2012). The relationship between alertness and executive control. J. Exp. Psychol. Hum. Percept. Perform. 38, 1530–1540. doi: 10.1037/a0027875, PMID: [DOI] [PubMed] [Google Scholar]
  96. Williams R. S., Biel A. L., Wegier P., Lapp L. K., Dyson B. J., Spaniol J. (2016). Age differences in the attention network test: evidence from behavior and event-related potentials. Brain Cogn. 102, 65–79. doi: 10.1016/j.bandc.2015.12.007, PMID: [DOI] [PubMed] [Google Scholar]
  97. Williams K. N., Kemper S. (2010). Interventions to reduce cognitive decline in aging. J. Psychosoc. Nurs. Ment. Health Serv. 48, 42–51. doi: 10.3928/02793695-20100331-03, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
  98. Wollman I., Arias P., Aucouturier J. J., Morillon B. (2020). Neural entrainment to music is sensitive to melodic spectral complexity. J. Neurophysiol. 123, 1063–1071. doi: 10.1152/jn.00758.2018, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
  99. Yin X., Zhao L., Xu J., Evans A. C., Fan L., Ge H., et al. (2012). Anatomical substrates of the alerting, orienting and executive control components of attention: focus on the posterior parietal lobe. PLoS One 7:e50590. doi: 10.1371/journal.pone.0050590, PMID: [DOI] [PMC free article] [PubMed] [Google Scholar]
  100. Zanto T. P., Gazzaley A. (2019). Aging of the frontal lobe. Handb. Clin. Neurol. 163, 369–389. doi: 10.1016/B978-0-12-804281-6.00020-3 [DOI] [PubMed] [Google Scholar]
  101. Zentner M., Grandjean D., Scherer K. R. (2008). Emotions evoked by the sound of music: characterization, classification, and measurement. Emotion 8, 494–521. doi: 10.1037/1528-3542.8.4.494, PMID: [DOI] [PubMed] [Google Scholar]
  102. Zhou S. S., Fan J., Lee T. M., Wang C. Q., Wang K. (2011). Age-related differences in attentional networks of alerting and executive control in young, middle-aged, and older Chinese adults. Brain Cogn. 75, 205–210. doi: 10.1016/j.bandc.2010.12.003, PMID: [DOI] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Data Availability Statement

The datasets presented in this study can be found in online repositories. The names of the repository/repositories and accession number(s) can be found at: https://osf.io/gf3km/.


Articles from Frontiers in Psychology are provided here courtesy of Frontiers Media SA

RESOURCES