Abstract
While there is an increasing shift in cognitive science to study perception of naturalistic stimuli, this study extends this goal to naturalistic contexts by assessing physiological synchrony across audience members in a concert setting. Cardiorespiratory, skin conductance, and facial muscle responses were measured from participants attending live string quintet performances of full-length works from Viennese Classical, Contemporary, and Romantic styles. The concert was repeated on three consecutive days with different audiences. Using inter-subject correlation (ISC) to identify reliable responses to music, we found that highly correlated responses depicted typical signatures of physiological arousal. By relating physiological ISC to quantitative values of music features, logistic regressions revealed that high physiological synchrony was consistently predicted by faster tempi (which had higher ratings of arousing emotions and engagement), but only in Classical and Romantic styles (rated as familiar) and not the Contemporary style (rated as unfamiliar). Additionally, highly synchronised responses across all three concert audiences occurred during important structural moments in the music—identified using music theoretical analysis—namely at transitional passages, boundaries, and phrase repetitions. Overall, our results show that specific music features induce similar physiological responses across audience members in a concert context, which are linked to arousal, engagement, and familiarity.
Subject terms: Neuroscience, Physiology, Psychology
Introduction
While there is an increasing shift in cognitive science to study human perception of naturalistic stimuli (e.g., real-world movies or music1), such research is still required in more naturalistic contexts. A concert setting provides one promising context to which research focusing on perception and experience of music can be extended; not only does it afford one possible naturalistic setting for music listening, but also live performances can evoke stronger emotional responses2–4 and offer a more immersive experience5,6. Although brain imaging techniques can implicitly measure undisturbed (i.e., without behavioural ratings) naturalistic musical perception as it evolves7–10, these methods lack applicability in a wider range of typical listening situations; therefore, more portable methods for measuring continuous responses such as motion capture11 or mobile measurement of peripheral physiology12,13 are required. As our interest lay in the musical experience within the context of a Western art music concert—in which listeners are typically still6,14—we focused on physiological responses of the autonomic system (ANS).
Previous research shows that certain physiological signatures indicative of a momentary ANS activations, regarding (phasic) skin conductance response (SCR, i.e., sweat secretion), heart rate (HR), and respiration rate (RR) as well as responses of facial muscles (electromyography [EMG] measurement), are related to orientation responses15,16 and affective processing17–20. Event-related changes in SCR, HR, RR, and EMG—reflecting a startle21–23 or orienting response24—have been associated with pitch changes25,26 and tone loudness (the louder the sound, the greater the SCR amplitude27,28), as well as deviations in timbre, rhythm, and tempo28. In other words, such signatures may be a response to novelty in stimuli.
Additionally, physiological responses have been shown to occur in response to arousing acoustic features17–20. For example, faster and increasing tempi are associated with greater arousal (reflecting emotions such as happiness) in the music29–32, which correspond to increased HR32–35, RR36, and SCR28,32,35,37,38 in a listener. Slower-paced music (reflecting low arousal emotions such as sadness) reduces HR, RR, and SCR36,39,40. Timbral features, such as brighter tones and higher spectral centroid, are associated with higher arousal29,41,42, which correlates somewhat to RR and SCR35,43,44. Loudness is positively correlated with arousal29,45,46, and, correspondingly, changes in SCR47,48 and HR29,49. Ambiguous harmony—operationalised as unexpected harmonic chords (i.e., out-of-key chords in place of tonic chords) and unpredictable notes (i.e., surprising notes within a predictable melodic sequence)—as well as dissonance may also be perceived as tense/arousing50,51 and lead to event-related increases in SCR50,52,53. Similarly, new or unprepared harmony, enharmonic changes, and harmonic acceleration to cadences have been found to evoke chills54, which are also related to increased SCR, HR, and RR when listening to music55–60.. Overall, this shows that peripheral physiology is related to arousing acoustic features, namely faster tempos, harmonic ambiguity, loudness, and (to some extent) timbral brightness. Importantly, it seems that increases in self-reported arousal and physiological measures are time-locked together in an event-related fashion13,61, suggesting that an increase in reported arousal is simultaneously reflected by increased ANS responses.
It is also worth noting that some ANS responses may be modulated by musical style. Previous studies found that HR increases with faster tempo in Classical music, but decreases with faster tempo in rock music34. HR is overall lower in atonal, compared to tonal music, independent of the emotional characteristics of the music62.
Although previous research generally supports the idea that specific physiological features are associated with specific musical features and styles, some of these studies (for reasons of experimental control) have carefully chosen and cut or constructed stimuli to have little variability in acoustic features (e.g., they use a constant tempo and normalise loudness). However, more research into full-length naturalistic stimuli—which typically have a rich dynamic variation of interdependent features—is required1.
While previous work using naturalistic music has correlated neural and physiological responses to dynamically changing acoustic features8,10,43,63, or extracted epochs based on information content in the music13, perhaps a more robust way to identify systematic responses to naturalistic stimuli is via synchrony of responses12,64–67, in particular via inter-subject correlation (ISC, see review68). This method—in which (neural) responses are correlated across participants exposed to naturalistic stimuli69—is based on the assumption that signals not related to processing stimuli would not be correlated. Such synchrony research has demonstrated that highly similar responses occur across subjects when exposed to naturalistic films69–73, spoken dialogue74–76 and text77,78, dance79, and music7–9,80, strongly suggesting that highly reliable and time-locked responses can be evoked by (seemingly uncontrolled) complex stimuli (for a review see81). Although ISC in functional magnetic resonance imaging (fMRI) studies can identify regions of interest (ROIs) for further analysis69, ISC can also assess which kind of feature(s) within dynamically evolving stimuli evoke highly correlated responses.
In response to auditory stimuli, higher synchronisation (operationalized via ISC) of participants’ responses has been associated with emotional arousal, structural coherence, and familiarity. In terms of emotional arousal, higher correlation coefficients of electroencephalography (EEG)71, SCR, and respiration82 (as well as higher hemodynamic activity in ROIs that had highly correlated activity between participants’ fMRI81) coincided with moments of high arousal in films, such as a close-up of a revolver71, gun-shots or explosions69, as well as close-ups of faces and emotional shakiness in voice82. Additionally, physiological synchrony between co-present audience members of movie viewings and theatre performance correlate with convergence of emotional responses and evaluation12,72. Regarding structural coherence, ISC is higher when listening to original, compared to phase-scrambled, versions of music7,80 and spoken text78. Cardiovascular and respiratory synchrony (calculated here by Generalised Partial Directed Coherence) was lower when audiences listened to music with complex auditory profiles64, further suggesting physiological synchrony may be linked to structural coherence. ISC may additionally reflect familiarity and engagement: it is higher when listening to familiar musical styles, compared to unfamiliar styles; though upon repeated presentation, ISC drops with repetitions of familiar (but not unfamiliar) musical style9. However, many of these findings come from neural and laboratory contexts with mostly general descriptions of the stimuli. Thus, further research is required to deepen our understanding of music and ISC by investigating ISC with more portable methods such as physiology and exploring which musical features—characterised by a more comprehensive analysis—can evoke synchronised physiological responses in more naturalistic listening situations.
The overarching aim of the current study was to explore which musical feature(s) evoke systematic physiological responses during undisturbed, naturalistic music listening in a concert context. Participants attended one of three chamber music concerts, with live performances of string quintets by Beethoven (1770–1827), Dean (1961–), and Brahms (1833–1897) (four movements each), showcasing different musical styles (Viennese Classical, Contemporary, and Romantic, respectively) with varying tempo, tonality, compositional structure, character, and timbre. Psychological (emotion and absorption) ratings were collected from a short (2-minute) questionnaire immediately after each movement, and familiarity ratings were collected after each piece. Continuous physiological responses were measured throughout the concert, from which SCR, HR, RR, and EMG activity of 98 participants (Concert 1 [C1]: 36, C2: 41, C3: 21) was extracted. For each audience, ISC was calculated over a sliding window (5 musical bars long) for each physiological measure, representing the degree of collective synchrony of physiological responses over the time-course of each musical stimulus (see Fig. 1a). High and low physiological synchrony were operationalised based on criteria from Dmochowksi et al.71. Windows containing ISC values in the highest 20th percentile represented high synchrony (HS) windows, while windows with correlation ISC values in the 20th percentile centred around r = 0 (that is, lowest correlation values) represented low synchrony (LS) windows.
To characterise the music, quantitative values of low- and medium-level features as well as detailed descriptions of high-level, structural features were obtained using inter-disciplinary approaches of acoustic signal and score-based analyses, respectively. Acoustic features most commonly associated with orientation and arousal physiological responses (as described above) were extracted. Instantaneous tempo was calculated using inter-onset intervals (IOIs) between each beat to represent the speed. Root mean square (RMS) energy was computationally extracted to represent loudness46 of the music. As the centroid of the spectral distribution of an acoustic signal has been shown to be a main contributor to perception of timbre83 and timbral brightness84,85, spectral centroid was computationally extracted (with higher values representing brighter timbre). Key clarity over time was calculated using an algorithm that correlates pitch class profiles (of the acoustic signal) with pre-defined key profiles (see86–88). The correlation coefficient associated with the best-matched key represents a quantitative measure of clarity of key, which is highly (positively) correlated with perceptual ratings of key clarity (i.e., the reverse of harmonic ambiguity)8. As we used naturalistic music, it was also important to consider stylistic and structural features of the music which cannot be analysed quantitatively (yet). Therefore, we prepared a detailed music theoretical analysis, indicating harmonic progressions, thematic and motivic relations, phrase rhythm, formal functions, and the overall repetition schema of the pieces89,90.
Based on the assumptions that ISC can identify systematic responses during naturalistic stimuli perception, and that certain physiological signatures and synchrony are related to musical features and self-reported states, the following hypotheses drove our research: (1) windows of high physiological synchrony—identified by ISC analysis—represent systematic physiological responses that are typically associated with event-related arousal responses; (2) highly correlated physiological responses during naturalistic music listening in a live concert context can be predicted by quantitative values of typically arousing acoustic features (higher RMS energy and spectral centroid, faster tempo, and lower key clarity), which may be modulated by the different styles. Additionally, we were (3) interested in the relations between ISC and higher-level musical features, such as compositional structure of the music. In light of the replicability crisis91, we tested whether robust physiological responses to music would be consistent across repeated concert performances. Such an approach simultaneously provides data regarding the stability of using a concert as an experimental setting.
Results
Acoustic comparisons: differences between performances and styles
Before testing our hypotheses, we assessed whether the extracted acoustic features were comparable across performances. As the music was performed by professional musicians who were instructed to play as similarly as possible across the concerts, statistical tests confirmed our expectation that all concert performances would be acoustically similar. No significant differences occurred between performances for loudness, tempo, timbre, and length, with Pearson correlations of instantaneous tempo, timbre, and loudness between all performances reaching r > 0.6, p < 0.001 (see Supplementary Tables S1a, S1b, and S1c). This confirms that performances were comparable enough to allow for further statistical comparisons of listeners’ physiology between audiences (i.e., that observed physiological responses are not attributable to unintended differences in the performances between concerts).
Because we hypothesized that style may play a role in driving physiological differences—and that certain styles are defined by differences in acoustic features—loudness, timbre, tempo, and key clarity were compared between styles. Contrasts revealed that Dean (Contemporary) had significantly lower RMS energy compared to Brahms (Romantic, p < 0.05). Dean also had significantly lower key clarity compared to both Beethoven (Classical, p < 0.012) and Brahms (p < 0.001), and significantly higher spectral centroid compared to Beethoven (p < 0.037) and Brahms (p < 0.046) (see Supplementary Figure S1 and Table S2a and S2b). Although tempo did not significantly differ between the styles (all p > 0.327), a division of tempi distribution in Beethoven and Brahms (Supplementary Figure S1) shows a typical composition practice of contrasting faster and slower movements in Classical and Romantic styles.
Summary
These acoustic checks confirm that our stimuli were comparable across concerts and that they offered a rich variation of acoustic features within and between pieces.
Physiological responses during moments of high versus low synchrony
Hypothesis 1 was assessed by comparing physiology in HS and LS windows. As shown in Fig. 1b, HR and RR were overall lower in HS compared to LS windows, confirmed by significant main effect of Synchrony for HR in all concerts and for RR in C3. Significant Synchrony by Bar interactions and contrasts for RR in C2 and C3 suggested that breathing accelerates from the onset bar (bar0) to the last bar (bar4) of the HS window (see Tables 1, 2). An HR increase also seemed to occur (see Fig. 1b), but did not withstand Bonferroni correction.
Table 1.
Concert | df | HR | RR | SCR | EMG | ||||
---|---|---|---|---|---|---|---|---|---|
F | p | F | p | F | p | F | p | ||
C1 | |||||||||
Synchrony | 1 | 13.316 | < 0.001 | 5.146 | 0.023 | 103.224 | < 0.001 | 20.044 | < 0.001 |
Bar | 4 | 4.138 | < 0.001 | 5.485 | < 0.001 | 28.969 | < 0.001 | 5.335 | < 0.001 |
Synch. × Bar | 4 | 3.336 | 0.010 | 2.190 | 0.068 | 26.392 | < 0.001 | 4.760 | < 0.001 |
C2 | |||||||||
Synchrony | 1 | 24.426 | < 0.001 | 8.045 | 0.005 | 52.673 | < 0.001 | 12.815 | < 0.001 |
Bar | 4 | 1.743 | 0.137 | 10.263 | < 0.001 | 26.981 | < 0.001 | 8.042 | < 0.001 |
Synch. × Bar | 4 | 0.846 | 0.495 | 6.300 | < 0.001 | 19.372 | < 0.001 | 5.974 | < 0.001 |
C3 | |||||||||
Synchrony | 1 | 10.065 | 0.002 | 25.527 | > 0.001 | 90.166 | < 0.001 | 29.240 | < 0.001 |
Bar | 4 | 3.739 | 0.005 | 10.696 | > 0.001 | 24.780 | < 0.001 | 6.405 | < 0.001 |
Synch. × Bar | 4 | 3.041 | 0.016 | 10.354 | > 0.001 | 15.977 | < 0.001 | 7.846 | < 0.001 |
Bonferroni-corrected threshold for significant effect was 0.05/12 = 0.004. Values highlighted in bold are statistically significant after Bonferroni corrections for multiple comparisons.
Table 2.
Pairwise comparison | HR | RR | ||||||||
---|---|---|---|---|---|---|---|---|---|---|
Estimate | SE | df | t | p | Estimate | SE | df | t | p | |
C1 | ||||||||||
HS–LS | − 0.036 | 0.007 | 4220 | − 5.073 | < 0.001 | − 0.010 | 0.007 | 4000 | − 1.504 | 0.133 |
HS: bar0–bar4 | − 0.052 | 0.016 | 4220 | − 3.315 | 0.042 | − 0.057 | 0.014 | 4000 | − 3.970 | 0.003 |
LS: bar0–bar4 | 0.016 | 0.016 | 4220 | 0.997 | 1.00 | − 0.004 | 0.015 | 4000 | − 0.273 | 1.00 |
C2 | ||||||||||
HS–LS | − 0.065 | 0.007 | 4225 | − 9.676 | < 0.001 | 0.002 | 0.006 | 3705 | − 0.256 | 0.798 |
HS: bar0–bar4 | − 0.040 | 0.015 | 4225 | − 2.415 | 0.710 | − 0.069 | 0.013 | 3705 | − 5.256 | < 0.001 |
LS: bar0–bar4 | − 0.003 | 0.015 | 4225 | − 0.193 | 1.0000 | 0.013 | 0.015 | 3705 | 0.837 | 0.998 |
C3 | ||||||||||
HS–LS | − 0.028 | 0.009 | 4195 | − 3.197 | 0.001 | − 0.028 | 0.008 | 3810 | − 3.518 | < 0.001 |
HS: bar0–bar4 | − 0.062 | 0.012 | 4195 | − 3.217 | 0.060 | − 0.091 | 0.017 | 3810 | − 5.495 | < 0.001 |
LS: bar0–bar4 | 0.0170 | 0.019 | 4195 | 0.880 | 1.00 | 0.041 | 0.018 | 3810 | 2.248 | 0.424 |
Pairwise comparison | SCR | EMG | ||||||||
---|---|---|---|---|---|---|---|---|---|---|
Estimate | SE | df | t | p | Estimate | SE | df | t | p | |
C1 | ||||||||||
HS–LS | 0.071 | 0.009 | 4220 | 8.364 | < 0.001 | 0.014 | 0.006 | 4220 | 2.505 | 0.012 |
HS: bar0–bar4 | 0.174 | 0.019 | 4220 | 9.161 | < 0.001 | 0.034 | 0.013 | 4220 | 2.683 | 0.33 |
LS: bar0–bar4 | − 0.062 | 0.019 | 4220 | − 3.252 | 0.052 | − 0.010 | 0.013 | 4220 | − 0.780 | 1.00 |
C2 | ||||||||||
HS–LS | 0.025 | 0.007 | 4200 | 3.428 | < 0.001 | − 0.001 | 0.006 | 4215 | − 0.084 | 0.933 |
HS: bar0–bar4 | 0.148 | 0.017 | 4200 | 8.924 | < 0.001 | 0.054 | 0.012 | 4215 | 4.391 | < 0.001 |
LS: bar0–bar4 | − 0.028 | 0.016 | 4200 | − 1.682 | 1.00 | − 0.010 | 0.012 | 4215 | − 0.845 | 1.00 |
C3 | ||||||||||
HS–LS | 0.095 | 0.010 | 4205 | 9.796 | < 0.001 | 0.018 | 0.007 | 4215 | 2.504 | 0.012 |
HS: bar0–bar4 | 0.185 | 0.022 | 4205 | 8.530 | < 0.001 | 0.061 | 0.016 | 4215 | 3.758 | 0.008 |
LS: bar0–bar4 | − 0.025 | 0.022 | 4205 | − 1.174 | 1.00 | − 0.030 | 0.016 | 4215 | − 1.868 | 1.00 |
HS–LS denotes the overall difference between HS and LS windows. Bar0–bar4 denotes the difference between the beginning and the end of the window separately in HS and LS windows. Contrasts (Bonferroni adjusted) were calculated with emmeans package in R. Values highlighted in bold are statistically significant after Bonferroni corrections for multiple comparisons.
SCR and EMG activity were overall higher in HS windows, compared to LS windows, confirmed by significant main effect of synchrony for SCR and EMG in all three concerts. Significant Synchrony by Bar interactions and contrasts suggested that SCR (all concerts) and EMG activity (C2 and C3) decreased across HS windows (bar0–bar4, see Tables 1, 2). Looking at a wider range of 4 bars before and after the correlation window, it seems that ISC identifies the second half of an EMG peak amplitude and SCR peak, i.e., a momentary increase of sweat secretion (see Fig. 1b).
Summary
Compared to the LS moments, HS windows contained higher SCR and EMG activity, and increasing RR. Such responses correspond to a momentary activation of the sympathetic division of the ANS, and have also been associated with self-reported arousal17–20. These results support our first hypothesis that ISC can identify systematic, event-related physiological responses, indicative of increased arousal.
Acoustic properties as predictors of audience synchrony
Regarding our second hypothesis, quantitative values of tempo, key clarity, spectral centroid, and RMS during high and low synchrony windows were compared using logistic regression.
Single physiology measures
Tempo significantly predicted synchronised RR arousal responses in Beethoven and Brahms C1 and C2, and significantly predicted synchronised SCR arousal responses in Beethoven C2 and Brahms all concerts, where faster tempi increased the probability of synchronised RR and SCR responses across audience members (see Table 3 and Fig. 2). Slower tempi significantly increased probability of synchronised HR arousal responses in Beethoven, but only for C2 (see Table 3). Higher RMS energy significantly increased the probability of RR synchrony in Dean C1 and SCR synchrony in Beethoven C1 (see Table 3). No significant results occurred either for EMG synchrony or for spectral centroid or key clarity.
Table 3.
Beethoven | Dean | Brahms | |||||||
---|---|---|---|---|---|---|---|---|---|
Estimate | SE | p value | Estimate | SE | p value | Estimate | SE | p value | |
Respiration rate | |||||||||
C1 | |||||||||
(Intercept) | − 1.211 | 0.897 | 0.177 | 1.059 | 2.150 | 0.6222 | − 0.728 | 1.255 | 0.562 |
Tempo | 0.0110 | 0.003 | > 0.001 | 0.003 | 0.009 | 0.707 | 0.018 | 0.005 | > 0.001 |
RMS | − 22.457 | 15.574 | 0.149 | 212.796 | 50.117 | > 0.001 | − 6.563 | 15.577 | 0.673 |
S. centroid | 0.000 | 0.000 | 0.439 | − 0.000 | 0.006 | 0.559 | − 0.000 | 0.000 | 0.809 |
Key clarity | − 0.151 | 1.114 | 0.891 | − 5.708 | 2.109 | 0.007 | − 1.218 | 1.569 | 0.438 |
C2 | |||||||||
(Intercept) | − 0.343 | 1.034 | 0.740 | 2.376 | 1.898 | 0.211 | − 1.451 | 1.297 | 0.263 |
Tempo | 0.016 | 0.004 | > 0.001 | − 0.006 | 0.009 | 0.502 | 0.028 | 0.004 | > 0.001 |
RMS | 0.610 | 17.630 | 0.972 | 58.895 | 39.231 | 0.133 | − 17.878 | 17.983 | 0.320 |
S. centroid | − 0.001 | 0.000 | 0.047 | − 0.001 | 0.001 | 0.049 | − 0.000 | 0.000 | 0.495 |
Key clarity | − 0.371 | 1.111 | 0.738 | − 0.305 | 2.017 | 0.880 | − 1.089 | 1.713 | 0.525 |
C3 | |||||||||
(Intercept) | − 1.614 | 0.953 | 0.090 | − 0.320 | 1.521 | 0.833 | − 0.3972 | 1.134 | 0.726 |
Tempo | 0.007 | 0.003 | 0.016 | 0.004 | 0.007 | 0.537 | 0.008 | 0.003 | 0.0188 |
RMS | 23.926 | 15.481 | 0.122 | 33.370 | 34.602 | 0.335 | − 22.432 | 15.333 | 0.143 |
S. centroid | 0.000 | 0.000 | 0.618 | − 0.0002 | 0.001 | 0.628 | − 0.000 | 0.000 | 0.384 |
Key clarity | 0.947 | 1.187 | 0.425 | − 0.478 | 1.685 | 0.777 | 0.929 | 1.415 | 0.511 |
Beethoven | Dean | Brahms | |||||||
---|---|---|---|---|---|---|---|---|---|
Estimate | SE | p value | Estimate | SE | p value | Estimate | SE | p value | |
Skin conductance | |||||||||
C1 | |||||||||
(Intercept) | − 1.959 | 0.913 | 0.032 | 3.522 | 1.813 | 0.052 | − 5.518 | 1.334 | > 0.001 |
Tempo | 0.007 | 0.003 | 0.016 | − 0.015 | 0.008 | 0.0551 | 0.028 | 0.005 | > 0.001 |
RMS | 76.201 | 17.351 | > 0.001 | 6.880 | 36.553 | 0.851 | 22.717 | 15.291 | 0.137 |
S. centroid | 0.001 | 0.000 | 0.048 | 0.000 | 0.001 | 0.728 | 0.006 | 0.000 | 0.134 |
Key clarity | − 0.626 | 1.024 | 0.541 | − 4.674 | 1.795 | 0.009 | 1.6770 | 1.532 | 0.274 |
C2 | |||||||||
(Intercept) | − 1.015 | 0.913 | 0.266 | − 3.901 | 2.091 | 0.061 | − 5.567 | 1.470 | > 0.001 |
Tempo | 0.011 | 0.003 | > 0.001 | 0.010 | 0.009 | 0.264 | 0.036 | 0.006 | > 0.001 |
RMS | 17.223 | 15.542 | 0.268 | 29.043 | 43.571 | 0.505 | 44.689 | 16.671 | 0.007 |
S. centroid | − 0.000 | 0.000 | 0.877 | 0.002 | 0.0001 | 0.006 | 0.000 | 0.000 | 0.438 |
Key clarity | − 0.300 | 1.094 | 0.784 | − 1.475 | 2.094 | 0.481 | 0.302 | 1.740 | 0.862 |
C3 | |||||||||
(Intercept) | − 0.660 | 0.883 | 0.455 | 0.512 | 1.576 | 0.745 | − 4.412 | 1.257 | > 0.001 |
Tempo | 0.001 | 0.003 | 0.606 | 0.011 | 0.007 | 0.095 | 0.015 | 0.004 | > 0.001 |
RMS | 21.770 | 15.067 | 0.148 | − 73.883 | 40.619 | 0.069 | 4.975 | 16.431 | 0.762 |
S. centroid | 0.000 | 0.000 | 0.152 | − 0.001 | 0.001 | 0.174 | 0.001 | 0.000 | 0.209 |
Key clarity | − 0.338 | 1.016 | 0.740 | 0.253 | 1.714 | 0.882 | 2.825 | 1.485 | 0.057 |
Beethoven | Dean | Brahms | |||||||
---|---|---|---|---|---|---|---|---|---|
Estimate | SE | p value | Estimate | SE | p value | Estimate | SE | p value | |
Heart rate | |||||||||
C1 | |||||||||
(Intercept) | − 0.7443 | 0.916 | 0.416 | 5.459 | 2.028 | 0.007 | 0.996 | 1.180 | 0.399 |
Tempo | 0.002 | 0.003 | 0.566 | − 0.008 | 0.007 | 0.261 | − 0.003 | 0.003 | 0.343 |
RMS | − 25.496 | 16.908 | 0.132 | − 86.004 | 40.332 | 0.033 | 1.825 | 13.783 | 0.895 |
S. centroid | 0.000 | 0.0003 | 0.103 | − 0.001 | 0.001 | 0.048 | − 0.001 | 0.000 | 0.161 |
Key clarity | 0.008 | 0.979 | 0.993 | − 3.620 | 2.063 | 0.079 | − 0.036 | 1.532 | 0.981 |
C2 | |||||||||
(Intercept) | − 0.435 | 0.899 | 0.629 | 0.480 | 1.490 | 0.747 | 0.367 | 1.213 | 0.762 |
Tempo | − 0.009 | 0.003 | 0.001 | − 0.008 | 0.006 | 0.214 | − 0.003 | 0.003 | 0.406 |
RMS | 20.089 | 14.853 | 0.176 | 25.795 | 32.573 | 0.428 | − 18.912 | 16.274 | 0.245 |
S. centroid | 0.001 | 0.0003 | 0.009 | 0.0002 | 0.001 | 0.626 | 0.001 | 0.000 | 0.052 |
Key clarity | 0.520 | 1.011 | 0.607 | − 1.765 | 1.376 | 0.200 | − 1.901 | 1.499 | 0.205 |
C3 | |||||||||
(Intercept) | 2.544 | 0.907 | 0.005 | 1.229 | 1.726 | 0.477 | − 2.139 | 1.1802 | 0.070 |
Tempo | 0.002 | 0.003 | 0.434 | 0.011 | 0.007 | 0.094 | 0.008 | 0.003 | 0.01 |
RMS | − 47.392 | 15.412 | 0.002 | − 68.302 | 39.045 | 0.080 | 4.159 | 14.888 | 0.7800 |
S. centroid | − 0.001 | 0.0002 | 0.029 | 0.001 | 0.001 | 0.307 | 0.000 | 0.000 | 0.761 |
Key clarity | − 2.087 | 1.065 | 0.050 | − 0.798 | 1.947 | 0.682 | 1.457 | 1.433 | 0.309 |
Synchrony (HS = 1, LS = 0) was the dependent variable, and tempo, key clarity, loudness, and spectral centroid (s. centroid) from the HS and LS bars as continuous predictors. The Bonferroni-corrected critical p value is 05/36 = 0.0014. Values highlighted in bold are statistically significant after Bonferroni corrections for multiple comparisons.
Multiple physiological measures
As HR, RR, and SCR, are all responses of the ANS, we assessed which musical features predicted a unified ANS response (i.e., when all three physiological measures were in synchrony simultaneously). Synchrony of the ANS as one entity (HR-RR-SCR) was not possible to model as no LS moments were found in the Dean piece for C1. Splitting ANS responses into paired combinations (HR-RR, HR-SCR, RR-SCR) yielded HS and LS moments in all three styles in all three concerts, allowing for further modelling. Logistic regressions revealed that faster passages (around 120 bpm, see Fig. 2) significantly increased probability of combined RR-SCR synchrony in Brahms C1 and C2 (see Table 4).
Table 4.
Beethoven | Dean | Brahms | |||||||
---|---|---|---|---|---|---|---|---|---|
Estimate | SE | p value | Estimate | SE | p value | Estimate | SE | p value | |
C1 | |||||||||
(Intercept) | − 0.880 | 2.106 | 0.676 | 12.646 | 11.099 | 0.254 | − 6.255 | 4.129 | 0.130 |
Tempo | 0.014 | 0.007 | 0.039 | − 0.036 | 0.049 | 0.461 | 0.058 | 0.016 | 0.0004 |
RMS | 82.658 | 39.948 | 0.038 | 535.054 | 277.231 | 0.054 | − 7.109 | 36.676 | 0.8463 |
S. centroid | 0.000 | 0.001 | 0.600 | − 0.001 | 0.002 | 0.508 | 0.001 | 0.002 | 0.323 |
Key clarity | − 2.644 | 2.559 | 0.302 | − 18.948 | 9.346 | 0.043 | − 2.956 | 3.566 | 0.407 |
C2 | |||||||||
(Intercept) | 0.935 | 2.486 | 0.707 | − 5.708 | 7.048 | 0.418 | − 13.149 | 7.235 | 0.069 |
Tempo | 0.014 | 0.006 | 0.030 | 0.006 | 0.034 | 0.859 | 0.211 | 0.067 | 0.002 |
RMS | 26.386 | 42.095 | 0.530 | 145.701 | 151.755 | 0.337 | 37.250 | 60.587 | 0.538 |
S. centroid | 0.000 | 0.001 | 0.759 | 0.002 | 0.0028 | 0.549 | − 0.001 | 0.002 | 0.407 |
Key clarity | − 4.533 | 3.077 | 0.141 | 3.612 | 6.376 | 0.571 | − 16.208 | 9.469 | 0.087 |
C3 | |||||||||
(Intercept) | − 9.744 | 3.155 | 0.002 | 27.244 | 22.265 | 0.221 | − 4.988 | 3.006 | 0.097 |
Tempo | 0.027 | 0.011 | 0.013 | 0.001 | 0.0212 | 0.943 | 0.025 | 0.011 | 0.031 |
RMS | 36.114 | 37.380 | 0.334 | − 193.756 | 194.915 | 0.320 | 27.009 | 42.694 | 0.527 |
S. centroid | 0.002 | 0.001 | 0.010 | − 0.0146 | 0.011 | 0.174 | 0.000 | 0.001 | 0.904 |
Key clarity | 4.609 | 3.127 | 0.140 | − 11.609 | 12.505 | 0.353 | 2.445 | 2.863 | 0.393 |
Synchrony (HS = 1, LS = 0) was the dependent variable, and tempo, key clarity, loudness, and spectral centroid (s. centroid) from the HS and LS bars as continuous predictors. The Bonferroni-corrected critical p value is 0.05/27 = 0.0019. Values highlighted in bold are statistically significant after Bonferroni corrections for multiple comparisons
Summary
Although HR, RR, and SCR synchrony were predicted by tempo, RMS, and spectral centroid, the only result that remained consistent across at least two concerts was that synchrony of RR (in Beethoven and Brahms), SCR, and RR-SCR (in Brahms) were predicted by faster tempi. These findings partially support our second hypothesis, showing that one typically arousing music feature (faster tempo) increased the probability of two synchronised arousal-related responses (higher SCR amplitude and increasing RR) consistently, an effect which was modulated by style. Our hypothesis that louder RMS, brighter timbre and lower key clarity would predict physiological synchrony was not supported by the current results.
Relationship between tempo and subjective experience
As we found a link between faster tempi and the highly synchronised SCR and RR increase (i.e., typical arousal responses) we wanted to further validate whether tempo was indeed related to self-reported emotion and engagement. Therefore, we correlated the mean tempo per movement with the psychological self-report data. Pearson correlations revealed that tempo positively correlated with engagement (e.g., feeling absorbed, concentrated), r = 0.475, p = 0.003, and positive high arousal emotions (e.g., energetic, joyful), r = 0.513, p = 0.001, with a medium and strong effect92, respectively.
Because the significance of our tempo results varied as a function of musical styles, we explored with descriptive statistics the possibility that familiarity with the piece might play a modulatory role. Beethoven and Brahms pieces were somewhat familiar to over a third of participants (41% and 35%, respectively), while the Dean piece was only somewhat familiar to 2.3% of participants (see Supplementary Table S3), suggesting that one interpretation why tempo significantly predicted physiological synchrony in the Beethoven and Brahms, but not Dean, could be due to differences in familiarity with the music.
Physiology synchrony across concerts in relation to the music theoretical analysis
With regard to our third hypothesis, we prepared a detailed music theoretical analysis of all the music (see Supplementary Table S4a, S4b, and S4c). We investigated musical events based on this analysis using moments of ‘salient’ physiological synchrony to find time points of interest in the music. Salient physiological responses were operationalised by two criteria: when (1) high physiological synchrony in any of the physiological measures occurred in all three concert audiences and (2) sustained synchrony occurred for more than one bar.
Overall, audience physiology seemed to synchronise around three types of musical events: (a) transitional passages with developing character, (b) clear boundaries between formal sections, and (c) phrase repetitions; all listed with descriptions in Supplementary Tables S5a, S5b, and S5c. Salient responses occurred during calming or arousing transitional passages (calming: Beethoven 1st movement, [Beethoven1], bars [b] 85–88; b287–293; Dean2, b70–71; Dean4, b75–77; Brahms4, b75–76; arousing: Beethoven1, b303–307; Dean2, b23–25; Brahms3, b6–8), characterised by a decrease (for calming passages) or an increase (for arousal passages) of loudness, texture, and pitch register. Other salient responses occurred when there was a clear boundary between functional sections, indicated through parameters such as a key change (e.g., between major and minor key in Beethoven3, b84–88; Brahms3, b58–61), a tempo change (e.g., Beethoven1 b328–331; Brahms4, b248–250), or a short silence (e.g., Beethoven1, b96–97; Beethoven4, b10–14). Lastly, salient responses occurred when a short phrase or motive was immediately repeated in a varied form, for example in an unexpected key, (e.g., Beethoven1, b35–37), elongated or truncated (e.g., Beethoven1, b85–88; 291–293), or with a different texture or pitch register (Brahms1, b90–91; Brahms3, b170–171). Since the immediate varied repetition of a short phrase is very common in Classical and Romantic styles, salient responses were also evoked when a phrase repetition occurred simultaneously with a transition or clear boundary (e.g., Beethoven1, b24–30; b136–138). With regard to style, the three categories are in line with the compositional conventions of the respective works: salient responses were found more often during transitions in the Romantic and Contemporary works, and during phrase repetitions and boundaries in the Classical work.
Discussion
This study assessed physiological aspects of the continuous music listening experience in a naturalistic environment using physiological ISC and inter-disciplinary stimulus analyses. We measured physiological responses of audiences listening to live instrumental music performances and examined which musical features evoked systematic physiological responses (i.e., synchronised responses, operationalised via ISC). Consistency of effects was assessed by repeating the same concert three times with different audiences. Importantly, no significant differences of length, loudness, tempo, or timbre across the concert performances were found, allowing us to assume that the musical stimuli were comparable across concerts.
Since previous research has identified typical physiological signatures as indices for felt arousal when listening to music, and since ISC is used to identify reliable responses across several individuals, we firstly hypothesised that ISC would identify similar physiological signatures of arousal. Our results supported this hypothesis: windows of high (compared to low) physiological synchrony contained significantly higher SCR/EMG and increasing RR, depicting similar physiological responses that have been previously related to self-reported arousal/tension evoked by music features like faster tempi28,32,35,37,38, peak loudness61 as well as unexpected harmonic chords51,93 or notes13 in music. These patterns of physiological responses indicate that windows of high physiology ISC, therefore, likely correspond to event-related moments of increased arousal.
Our second hypothesis—that low- and medium-level acoustic features can predict high physiological synchrony—was partially supported. Specifically, logistic regressions revealed that one typically arousing musical feature—faster tempi—consistently (i.e., in at least two concerts) predicted synchronised RR and SCR arousal responses. Additionally, tempo consistently predicted a more general ANS response, that is, when both RR and SCR of audience members became synchronised simultaneously. In line with previous work showing that tempo and rhythm are the most important musical features in determining physiological responses32, these findings suggest that tempo induces reliable physiological changes.
As faster tempo is typically perceived as more arousing29–32, our result that faster music increased probability of high physiological synchrony supports previous research linking high ISC to increased arousal69,71,82. Our findings further support the idea that ISC is related to stimulus engagement9: as slower music increases mind-wandering94, slower tempi may result in reduced attention to the music, leading to greater individual variability in physiological responses and subsequently lower ISC69,78. These connections between faster music with high physiological synchrony due to arousal and engagement were further confirmed by the psychological data, where faster tempi were positively correlated with audience members’ self-reported engagement (e.g., high concentration and feelings of being absorbed) and high arousal positive emotions (e.g., feeling energetic or joyful). It is important to note that faster tempi (centred around 120 bpm/2 Hz in the current study) seem to be a more physiologically optimal range for entrainment to music (see95 for review). As entrainment is difficult at note rates under 1 Hz (at least for non-expert musicians96), this may explain why synchrony did not occur in slower tempi of the current musical stimuli (which was centred around 50 bpm/0.83 Hz). It is therefore reasonable to speculate that entrainment, or the adaptation of autonomic physiological measures towards the musical tempo, might be a mechanism through which faster or optimally resonant tempi induce more-similar audience responses. In summary, we postulate that physiological synchrony may be predicted by faster music because of the physiological properties of the nervous system, which, when optimally driven, induce specific psychological experiences such as increased arousal and engagement.
It is of further interest that SCR and RR synchrony were more probable at faster tempi in the Classical and Romantic styles, but not in the Contemporary style, supporting previous studies where the same features evoke different physiological responses based on the style34,62. As Beethoven and Brahms were rated as more familiar compared to Dean, ISC differences between styles in the current results support a recent study showing that ISC is linked to familiarity with musical style9. Additionally, Beethoven and Brahms have relatively stable meters (i.e., very few instantaneous tempo changes within movements), whereas many Dean passages contained unpredictable meters (e.g., the first movement has alternating bars of four, five, or six beats per bar) and frequent tempo changes within movements. In view of this, the fact that synchrony probability changes between styles could be due to stimulus coherence, corroborating studies showing that higher ISC occurs in more predictable contexts78 and lower ISC occurs in versions of music where the beat is disrupted80. However, since we presented only one work per style (and tempo changes were not evenly represented across these styles)—a compromise dictated by the constraints of a naturalistic concert setting, complimentary research using a wider range of pieces (in different styles) is required to further assess the effects of coherence and familiarity on synchrony.
Although we show that synchrony was predicted by tempo (depending on the style), the hypothesis that louder RMS, brighter timbre, and lower key clarity would predict physiological synchrony was not met. This was unexpected, as orienting/startle response research consistently shows that loudness evokes highly replicable physiological responses in a controlled tone sequence24,27,28. Because loudness in the current study was embedded in naturalistic music, our findings highlight the generalisability limitations of reductionist stimuli to real-world contexts1. Indeed, previous work has shown that environmental sounds and music evoke different physiological responses; an increase in HR (index of a startle response23) occurred with arousing noises (e.g., a ringing telephone or storm), but not with music97. However, as we used such naturalistic music, it was important to not only explore quantitatively extracted low- and medium-level features, but also higher-level parameters in the music.
With regard to the hypothesis that synchrony corresponds to high-level/structural moments in the music, we observed that (from all moments in the music) transitional passages, clear boundaries, and immediate phrase repetitions in the music—identified using music theoretical analysis—coincided with highly synchronised physiological responses across concerts. Similar high-level features have been associated with physical responses such as shivers, laughter or tears (see Table IV in 54) or specifically chills61, which also typically correspond with increased physiological arousal55–60. However, none of the previous studies have derived these categories based on a detailed analysis of complete pieces and in the context of longer time spans.
In the current study, synchronised physiological responses occurred during arousing/calming transitional passages (characterised by changes in loudness, pitch register, and musical texture) and boundaries indicated by sudden tempo or key changes. Previous research has shown that unexpected musical events embedded in a predictable context may be perceived as arousing13,51,93, further corroborating findings that high ISC occurs at arousing moments69,71,82. Our results additionally align with the notion that audience members collectively ‘grip on’ to loudness and texture changes5. The finding that physiological synchrony occurred during tempo changes, supports the fact that disruptions of temporal expectations affect ANS responses28,98 and EEG synchrony80, where surprising events phase-reset ongoing physiological oscillations (see99,100 for reviews), leading—at least briefly—to an increase in audience synchrony around moments of phase resetting.
We also found that momentary synchronised responses occurred during short phrases which were immediately repeated in a varied form, hinting at a general attention towards repetitions in music101 and a recognition of thematic connections over longer time spans101. Our analysis further suggests that an interplay of various musical features—in addition to the simple repetition—increase attention, and subsequently synchrony, of all audience members to these musical moments. For instance, high audience synchrony occurred in some of the structurally most important moments of Beethoven’s 1st movement, where phrase repetition occurs simultaneously with a boundary (at the end of the exposition and with references to the primary theme at the end of the movement), and a boundary occurs simultaneously with transitional passages (deferred cadences; declined structural closures: e.g., b96–97, b301–302). The fact that high ISC occurred at phrase repetitions with an added novelty in the phrase (e.g., in a different key)—and also at important structural locations—not only supports research where novelty in stimuli evokes a physiological response16,18,21–28, but is also in line with music compositional practices, in which a composer tries to vary and develop thematic material102 (with different textures and/or harmonies) to keep listeners’ interest.
In conclusion, by measuring continuous music listening experience in a naturalistic setting of a chamber music concert, we show that systematic synchronised physiological responses (corresponding to typical arousal responses) across audience members are predicted by tempo (depending on style) and are linked to structural transitions, boundaries, and phrase repetitions. Using naturalistic music in a concert environment is beneficial in that participants are likely to have more realistic and stronger responses2–4. However, as this benefit makes our findings specific to the music we have used, future research should assess whether the current findings related to musical features and style are replicated with different (styles of) music. Additionally, further questions remain regarding the concert setting itself; for example, whether physiological effects and subjective experiences change with/without visual movements of the performer(s), with varying programming orders, and in different performance spaces6,103. Exploring musical experiences from pre-recorded or live performances—with or without the co-presence of others—may prove an interesting future research direction, especially with regard to the COVID-19 pandemic and current transformations of the concert itself.
Method
Experimental procedures
Experimental procedures are identical to Merrill et al.104. All experimental procedures were approved by the Ethics Council of the Max Planck Society, and undertaken with written informed consent of each participant. All research was performed in accordance with the Declaration of Helsinki.
Participants
129 participants attended one of three evening concerts. Some data were lost due to technical issues of server and user failures (N = 31) during data acquisition, leaving data from 98 participants for analysis. Gender and age were similarly distributed across concerts (see Table 5). Most participants reported that their highest level of education was a university degree, a German high school degree, or completed professional training. Musical Sophistication—assessed with the General Music Sophistication and Emotions subscales from the German version of the Goldsmiths Music Sophistication Index105,106—was similarly moderate (see Appendix Table 3 in101) across the three audience groups (see Table 5). Most participants reported that they regularly attend classical concerts and opera.
Table 5.
Concert | Total N | Gender | Age | Gold-MSI: emotion | Gold-MSI: general |
---|---|---|---|---|---|
Mean score (SD) | Mean score (SD) | ||||
C1 | 36 | F = 15, M = 17, na = 4 | 50% < 50 years old | 33.53 (4.75) | 69.84 (22.01) |
C2 | 41 | F = 16, M = 17, na = 8 | 50% < 55 years old | 31.17 (6.85) | 71.61 (21.97) |
C3 | 21 | F = 9, M = 12 | 50% < 40 years old | 33.24 (5.45) | 70.76 (19.33) |
Participants were asked to report their age by selecting age group (within a 5-year range from 18 to 99, i.e., 18–22, 23–27, 28–32, etc.).
Concert context
Three evening concerts (starting at 19.30 and ending at approximately 21.45) took place in a hybrid performance hall purpose-built for empirical investigations (the ‘ArtLab’ of the Max Planck Institute for Empirical Aesthetics in Frankfurt am Main, Germany). Care was taken to keep parameters (e.g., timing, lighting, temperature) as similar as possible across concerts. Professional musicians performed string quintets in the following order: Ludwig van Beethoven, op. 104 in C minor (1817), Brett Dean, ‘Epitaphs’ (2010), and Johannes Brahms, op. 111 in G major (1890), with a 20-minute interval between Dean and Brahms. This program was chosen to represent a typical chamber music concert.
Procedure
Participants were invited to arrive either one or one and half hour(s) before the concert began for physiological measurement preparation. Physiology was measured with a portable recording and amplifying device (https://plux.info/12-biosignalsplux) for the whole concert at 1000 Hz. Continuous blood volume pulse (BVP) was measured using a plethysmograph clip; respiration data were measured using a respiration belt (wrapped snugly around the lower rib cage); skin conductance was measured with electrodes placed on index and middle fingers of the non-dominant hand; and facial muscle activity was measured using electromyography, with adhesive electrodes placed above the zygomaticus major (‘smiling’) muscle on the left side of the face and ground placed on the mastoid. After each of the 12 movements, a short (2-min) pause was taken for the participants to fill in two questionnaires. The first questionnaire measured state absorption: eight items (e.g., ‘I was completely absorbed by the music’, ‘My mind was wandering’) were rated on a 5-point Likert scale from ‘strongly disagree’ to ‘strongly agree’107. The second questionnaire measured intensity of felt emotions using the GEneva Music-Induced Affect Checklist (GEMIAC108), where intensity of 14 classes of feelings (e.g., energetic/lively, tense/uneasy, nostalgic/sentimental) were rated on a 5-point Likert scale from ‘not at all’ to ‘very much’ (intensely experienced). Items were presented in German using comparable adjectives from the German version of the AESTHEMOS109. After each piece, participants rated how familiar the piece was (‘Yes’, ‘No’, and ‘Not sure’).
Data analysis
Musical feature extraction
Instantaneous tempo was calculated using inter-onset intervals (IOIs) between each beat (where beats were manually annotated by tapping each beat using Sonic Visualiser102). These IOIs were then converted to beats per minute (bpm). All other features were computationally extracted using the MIRToolbox (version 1.7.2)110 in MATLAB 2018b. In order to capture musical features meaningfully, different time windows were used to extract certain musical features. For features that change quickly on a timeframe of a less than a second—i.e., RMS energy (related to loudness), spectral centroid, brightness, and roughness (related to timbre)—a time window of 25 ms with a 50% overlap was used, as is typical in the music information retrieval literature8,111. Other features that are more context-dependent (that develop over a longer time frame), such as key clarity (i.e., the reverse of harmonic ambiguity) require a longer time-window, and were extracted similar to previous studies that assess the same musical feature, that is using a 3 s window8,41 with 33% hop factor (overlap)112. As previous time-series analyses have parsed data into meaningful units of clause and sentence lengths74, and as we wanted to aligned responses across concerts, each feature was averaged into a meaningful and comparable music unit: a bar (American: measure, on average 2 s long). It is worth noting that acoustic features can be distinguished between compositional features and performance features113, where the former are represented in the musical score (such as harmony) while the latter include features that can change between performances, namely how loud and fast musicians may perform the music. Because key clarity is a compositional feature (i.e., does not change between performances), we had the same values across all three concert performances.
When checking for independence of features114, Pearson correlations revealed that RMS and roughness correlated highly as did brightness and spectral centroid (all r > 0.7, p > 0.0001) in all movements. As RMS and spectral centroid are features more commonly used, compared to roughness and brightness113,115, and spectral centroid seems to best represent timbre83,116 and brightness117 perception, we kept only key clarity, RMS, spectral centroid, and tempo for further analysis. To check performance feature similarity between concerts, features were compared with concert (C1, C2, C3) as the independent variable. Pearson correlations were used to assess similarity of acoustic features over time between concerts. Correlations were considered adequate if they met a large effect size of concert r > 0.592. To compare acoustic features per style, linear mixed models with fixed effect of the works (Beethoven, Brahms, and Dean) and random intercept of movement were constructed with each acoustic feature (as the dependent variables) per concert.
Physiology pre-processing
Physiological data were pre-processed and analysed in MATLAB 2018b. Data were cut per movement. Missing data (gaps of less than 50 ms) were interpolated at the original sampling rate. Fieldtrip118 was used to pre-process BVP, respiration, and EMG data. BVP data were band-pass filtered between 0.8 and 20 Hz (4th order, Butterworth) and demeaned per movement. Adjacent systolic peaks were detected to obtain inter-beat intervals (IBIs) and an additional filter was added to remove any IBIs that were shorter than 300 ms, longer than 2 s, or had a change of more than 20% between adjacent IBIs (typical features of incorrectly identified IBIs119). After visual inspection and artefact removal, IBIs were converted to continuous heart rate (HR) by interpolation. Respiration data were low-pass filtered (0.6 Hz, 6th order, Butterworth) and demeaned. Maximum peaks were located and respiration rate (RR) was inferred by the peak intervals. EMG activity was band-pass filtered (between 90 and 130 Hz, 4th order, Butterworth), demeaned, and the absolute value of the Hilbert transform of the filtered signal was extracted and smoothed. Skin conductance data were pre-processed using Ledalab15 and decomposed into phasic and tonic activity. As we were interested in event-related responses, only (phasic) skin conductance responses (SCR) were used in further analyses. All pre-processed physiological data (SCR, HR, RR, EMG) were resampled at 20Hz17, z-scored within participant and movement, and averaged into bins per bar.
ISC analysis
We calculated a time-series ISC based on Simony et al.78 by forming p × n matrices (one for each SCR, HR, RR, and EMG, and for each of the twelve movements per concert), where p is the physiological response for each participant over n time points (bars across the movements). Correlations were calculated over a sliding window 5-bars long (approximately 10 s; the average bar length across the whole concert was 2 s), shifting one bar at a time. Fisher’s r-to-z transformation was applied to correlation coefficients per subject, then averaged z values were inverse transformed back to r values. The first 5 bars and the last 5 bars of each movement were discarded to remove common physiological responses evoked by the onset/offset of music76. ISC values per movement were concatenated within each concert (2238 bars), giving four physiological ISC measures per concert. These ISC traces represent the similarity of the audience members’ physiological responses over time (see Fig. 1a).
Following Dmochowski et al.71, high and low synchrony were defined using 20th percentiles. Windows containing the highest 20th percentile of ISC values were categorised as high synchrony (HS) windows. Windows with values in the lowest 20th percentile of correlation values (i.e., ISC values within a 20th percentile where r was centred around zero) were categorised as low synchrony (LS) windows. To obtain instances of overall ANS synchrony (i.e., across multiple physiological measures simultaneously), we identified where HS/LS moments of one physiological measure coincided with another physiological measure. Physiological responses at points of HS and LS were compared using linear models with fixed effects Synchrony (high/low) and Bar (bars 0–4). To investigate whether acoustic features predicted synchrony (HS/LS) of physiological responses across audience members, tempo, RMS energy, key clarity, and spectral centroid in bars of HS/LS were recovered. By dummy-coding Synchrony as a binary variable (HS as 1; LS as 0) logistic regression models were constructed to predict Synchrony for each physiological measure (dependent variable) with continuous predictors of tempo, key clarity, loudness, and spectral centroid from the HS and LS bars (all features were included, as perceived expression in music tends to be determined by multiple musical features30). (N.B.: no random intercept of movement was included, because not all movements contained HS/LS epochs). As we expected style to modulate the effect of these acoustic features in predicting synchrony, models were run separately per piece and concert.
Subjective ratings
A factor analysis was conducted for all ratings from the absorption and GEMIAC scales. Five latent variables were identified: (1) Positive high arousal emotions (energetic, powerful, inspired, joyful, filled with wonder, enchanted); (2) Negative high arousal emotions (tense, agitated, negative loadings of liking and feeling relaxed); (3) Mixed valence, low arousal emotions (relaxed, nostalgic, melancholic, feelings of tenderness, moved); (4) Engagement (concentrated, forgetting, absorbed, liking, with negative loadings of bored, indifferent, and mind-wandering); and (5) Dissociation (forgetting being in a concert and forgetting surroundings); see104 for their exact loadings. We (Pearson) correlated participants’ factor scores with musical features.
Music theoretical analysis
The scores of all works were analysed according to widely used methods for the respective styles. Musical events were analysed on the beat level (harmonic changes, cadences, texture changes, motivic relation120) and grouped into larger sections (thematic relations, formal functions/action spaces, repetition schemata89,90). The performance recordings served as references for passages which could have been interpreted equivocally in the score. After the analysis, passages involving high physiological synchrony were marked. These musical features were compared with each other, categorized across styles, and finally reduced to three categories: (a) transitional passages, (b) clear boundaries between formal sections, and c) phrase repetitions.
Statistical analyses
Statistical analyses were conducted in R121. Pearson correlations were computed using corr.test in the psych package122 and adjusted for false discovery rate using the Benjamin-Hochberg procedure. Linear (mixed) models were constructed using the lme4 package123; p values were calculated with the lmerTest package124 and using the Anova function in the car package125. Contrasts were assessed with the emmeans function (emmeans package126). Logistic regression models were run using a general linear model with a logit link function. Significance thresholds for p values for ANOVA, contrasts, and logistic regression models were adjusted using Bonferroni corrections.
Supplementary Information
Acknowledgements
Many thanks go to staff at the ArtLab (Frankfurt am Main), in particular Jutta Toelle, Alexander Lindau, Patrick Ulrich, and Eike Walkenhorst who helped organise the concerts; to Claudia Lehr and Freya Materne for organizing the invitations of the participants, and finally, thanks to the many assistants during data collection, particularly Till Gerneth, Sandro Wiesmann, Nancy Schön, and Simone Franz. We would like to thank Folkert Uhde, as part of the Experimental Concert Research (ERC) project, who set up the musical program for this concert. Thanks also to Cornelius Abel and Alessandro Tavano for data quality checks; to Helen Singer for annotating musical beats.
Author contributions
A.C. design of the work; statistical and music theoretical analysis; data interpretation; writing—original draft and figures. L.K.F. design of the work; statistical analysis; data interpretation; writing—review and editing. L.T.F. music theoretical analysis; Supplementary Table S4; writing—review and editing. M.T. conception. M.W.F. conception; data acquisition; data interpretation; writing—review. J.M. design of the work; data acquisition and statistical analysis; data interpretation; writing—review and editing.
Funding
Open Access funding enabled and organized by Projekt DEAL.
Data availability
Data of this study are available from the corresponding author upon request.
Competing interests
The authors declare no competing interests.
Footnotes
Publisher's note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary Information
The online version contains supplementary material available at 10.1038/s41598-021-00492-3.
References
- 1.Nastase S, Goldstein A, Hasson U. Keep it real: Rethinking the primacy of experimental control in cognitive neuroscience. Neuroimage. 2020;222:117254. doi: 10.1016/j.neuroimage.2020.117254. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.Coutinho E, Scherer KR. The effect of context and audio–visual modality on emotions elicited by a musical performance. Psychol. Music. 2017;45:550–569. doi: 10.1177/0305735616670496. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3.Gabrielsson A, Lindström Wik S. Strong experiences related to music: A descriptive system. Music. Sci. 2003;7:157–217. [Google Scholar]
- 4.Lamont AM. University students’ strong experiences of music: Pleasure, engagement, and meaning. Music. Sci. 2011;15:229–249. [Google Scholar]
- 5.Phillips M, et al. What determines the perception of segmentation in contemporary music? Front. Psychol. 2020;11:1–14. doi: 10.3389/fpsyg.2020.01001. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6.Wald-Fuhrmann M, et al. Music listening in classical concerts: Theory, Literature review, and research program. Front. Psychol. 2021;12:1324. doi: 10.3389/fpsyg.2021.638783. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7.Abrams DA, et al. Inter-subject synchronization of brain responses during natural music listening. Eur. J. Neurosci. 2013;37:1458–1469. doi: 10.1111/ejn.12173. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Alluri V, et al. Large-scale brain networks emerge from dynamic processing of musical timbre, key and rhythm. Neuroimage. 2012;59:3677–3689. doi: 10.1016/j.neuroimage.2011.11.019. [DOI] [PubMed] [Google Scholar]
- 9.Madsen J, Margulis EH, Simchy-Gross R, Parra LC. Music synchronizes brainwaves across listeners with strong effects of repetition, familiarity and training. Sci. Rep. 2019;9:1–8. doi: 10.1038/s41598-019-40254-w. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Burunat I, et al. The reliability of continuous brain responses during naturalistic listening to music. Neuroimage. 2016;124:224–231. doi: 10.1016/j.neuroimage.2015.09.005. [DOI] [PubMed] [Google Scholar]
- 11.Swarbrick D, et al. How live music moves us: Head movement differences in audiences to live versus recorded music. Front. Psychol. 2019;9:1–11. doi: 10.3389/fpsyg.2018.02682. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.Ardizzi M, Calbi M, Tavaglione S, Umiltà MA, Gallese V. Audience spontaneous entrainment during the collective enjoyment of live performances: Physiological and behavioral measurements. Sci. Rep. 2020;10:1–12. doi: 10.1038/s41598-020-60832-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.Egermann H, Pearce MT, Wiggins GA, McAdams S. Probabilistic models of expectation violation predict psychophysiological emotional responses to live concert music. Cogn. Affect. Behav. Neurosci. 2013;13:533–553. doi: 10.3758/s13415-013-0161-y. [DOI] [PubMed] [Google Scholar]
- 14.Thorau C, Ziemer H. The Oxford Handbook of Music Listening in the 19th and 20th Centuries. Oxford University Press; 2019. [Google Scholar]
- 15.Benedek M, Kaernbach C. A continuous measure of phasic electrodermal activity. J. Neurosci. Methods. 2010;190:80–91. doi: 10.1016/j.jneumeth.2010.04.028. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16.Frith CD, Allen HA. The skin conductance orienting response as an index of attention. Biol. Psychol. 1983;17:27–39. doi: 10.1016/0301-0511(83)90064-9. [DOI] [PubMed] [Google Scholar]
- 17.Bradley MM, Lang P. Affective reactions to acoustic stimuli. Psychophysiology. 2000;37:204–215. [PubMed] [Google Scholar]
- 18.Bradley MM. Natural selective attention: Orienting and emotion. Psychophysiology. 2009;46:1–11. doi: 10.1111/j.1469-8986.2008.00702.x. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19.Cacioppo JT, Berntson G, Larsen J, Poehlmann K, Ito T. The psychophysiology of emotions. In: Lewis R, Haviland-Jones JM, editors. Handbook of Emotions. Guilford Press; 2000. pp. 173–191. [Google Scholar]
- 20.Hodges DA. Psychophysiological measures. In: Juslin PN, Sloboda JA, editors. Handbook of Music and Emotion: Theory, Research, Applications. Oxford University Press; 2011. pp. 279–311. [Google Scholar]
- 21.Brown P, et al. New observations on the normal auditory startle reflex in man. Brain. 1991;114:1891–1902. doi: 10.1093/brain/114.4.1891. [DOI] [PubMed] [Google Scholar]
- 22.Dimberg U. Facial electromyographic reactions and autonomic activity to auditory stimuli. Biol. Psychol. 1990;31:137–147. doi: 10.1016/0301-0511(90)90013-m. [DOI] [PubMed] [Google Scholar]
- 23.Graham FK, Clifton RK. Heart-rate change as a component of the orienting response. Psychol. Bull. 1966;65:305–320. doi: 10.1037/h0023258. [DOI] [PubMed] [Google Scholar]
- 24.Barry RJ, Sokolov EN. Habituation of phasic and tonic components of the orienting reflex. Int. J. Psychophysiol. 1993;15:39–42. doi: 10.1016/0167-8760(93)90093-5. [DOI] [PubMed] [Google Scholar]
- 25.Lyytinen H, Blomberg A-P, Näätänen R. Event-related potentials and autonomic responses to a change in unattended auditory stimuli. Psychophysiology. 1992;29:523–534. doi: 10.1111/j.1469-8986.1992.tb02025.x. [DOI] [PubMed] [Google Scholar]
- 26.Sidle DA, Heron PA. Effects of length of training and amount of tone frequency change on amplitude of autonomic components of the orienting response. Psychophysiology. 1976;13:281–287. doi: 10.1111/j.1469-8986.1976.tb03076.x. [DOI] [PubMed] [Google Scholar]
- 27.Barry RJ. Low-intensity auditory stimulation and the GSR orienting response. Physiol. Psychol. 1975;3:98–100. [Google Scholar]
- 28.Chuen L, Sears D, McAdams S. Psychophysiological responses to auditory change. Psychophysiology. 2016;53:891–904. doi: 10.1111/psyp.12633. [DOI] [PubMed] [Google Scholar]
- 29.Coutinho E, Cangelosi A. Musical emotions: Predicting second-by-second subjective feelings of emotion from low-level psychoacoustic features and physiological measurements. Emotion. 2011;11:921–937. doi: 10.1037/a0024700. [DOI] [PubMed] [Google Scholar]
- 30.Gabrielsson A. The relationship between musical structure and perceived expression. In: Hallam S, Cross I, Thaut M, editors. The Oxford Handbook of Music Psychology. Oxford University Press; 2008. pp. 141–150. [Google Scholar]
- 31.Gabrielsson A, Juslin PN. Emotional expression in music performance between the performer’s intention and the listener’s experience. Psychol. Music. 1996;24:68–91. [Google Scholar]
- 32.Gomez P, Danuser B. Relationships between musical structure and psychophysiological measures of emotion. Emotion. 2007;7:377–387. doi: 10.1037/1528-3542.7.2.377. [DOI] [PubMed] [Google Scholar]
- 33.Bernardi L, Porta C, Sleight P. Cardiovascular, cerebrovascular, and respiratory changes induced by different types of music in musicians and non-musicians: The importance of silence. Heart. 2006;92:445–452. doi: 10.1136/hrt.2005.064600. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 34.Dillman Carpentier FR, Potter RF. Effects of music on physiological arousal: Explorations into tempo and genre. Media Psychol. 2007;10:339–363. [Google Scholar]
- 35.Egermann H, Fernando N, Chuen L, McAdams S. Music induces universal emotion-related psychophysiological responses: Comparing Canadian listeners to Congolese Pygmies. Front. Psychol. 2015;5:1341. doi: 10.3389/fpsyg.2014.01341. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 36.Krumhansl CL. An exploratory study of musical emotions and psychophysiology. Can. J. Exp. Psychol. Can. Psychol. Exp. 1997;51:336–353. doi: 10.1037/1196-1961.51.4.336. [DOI] [PubMed] [Google Scholar]
- 37.Khalfa S, Peretz I, Jean-Pierre B, Manon R. Event-related skin conductance responses to musical emotions in humans. Neurosci. Lett. 2002;328:145–149. doi: 10.1016/s0304-3940(02)00462-7. [DOI] [PubMed] [Google Scholar]
- 38.van der Zwaag M, Westerink JHDM, van den Broek EL. Emotional and psychophysiological responses to tempo, mode, and percussiveness. Music. Sci. 2011;15:250–269. [Google Scholar]
- 39.Etzel JA, Johnsen EL, Dickerson J, Tranel D, Adolphs R. Cardiovascular and respiratory responses during musical mood induction. Int. J. Psychophysiol. 2006;61:57–69. doi: 10.1016/j.ijpsycho.2005.10.025. [DOI] [PubMed] [Google Scholar]
- 40.Gupta U, Gupta BS. Psychophysiological reactions to music in male coronary patients and healthy controls. Psychol. Music. 2015;43:736–755. [Google Scholar]
- 41.Brattico E, et al. A functional MRI study of happy and sad emotions in music with and without lyrics. Front. Psychol. 2011;2:1–16. doi: 10.3389/fpsyg.2011.00308. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 42.Gingras B, Marin MM, Fitch WT. Beyond intensity: Spectral features effectively predict music-induced subjective arousal. Q. J. Exp. Psychol. 2014;67:1428–1446. doi: 10.1080/17470218.2013.863954. [DOI] [PubMed] [Google Scholar]
- 43.Bannister S, Eerola T. Suppressing the chills: Effects of musical manipulation on the chills response. Front. Psychol. 2018;9:2046. doi: 10.3389/fpsyg.2018.02046. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 44.Gorzelanczyk EJ, Podlipniak P, Walecki P, Karpinski M, Tarnowska E. Pitch syntax violations are linked to greater skin conductance changes, relative to timbral violations—The predictive role of the reward system in perspective of cortico-subcortical loops. Front. Psychol. 2017;8:1–11. doi: 10.3389/fpsyg.2017.00586. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 45.Juslin PN, Laukka P. Communication of emotions in vocal expression and music performance: Different channels, same code? Psychol. Bull. 2003;129:770–814. doi: 10.1037/0033-2909.129.5.770. [DOI] [PubMed] [Google Scholar]
- 46.Laurier, C., Lartillot, O., Eerola, T. & Toiviainen, P. Exploring relationships between audio features and emotion in music. In Triennial Conference of European Society for the Cognitive Sciences of Music 260–264 (2009). 10.3389/conf.neuro.09.2009.02.033.
- 47.Bach DR, Neuhoff JG, Perrig W, Seifritz E. Looming sounds as warning signals: The function of motion cues. Int. J. Psychophysiol. 2009;74:28–33. doi: 10.1016/j.ijpsycho.2009.06.004. [DOI] [PubMed] [Google Scholar]
- 48.Olsen KN, Stevens CJ. Psychophysiological response to acoustic intensity change in a musical chord. J. Psychophysiol. 2013;27:16–26. [Google Scholar]
- 49.Wilson CV, Aiken LS. The effect of intensity levels upon physiological and subjective affective response to rock music. J. Music Ther. 1977;14:60–76. [Google Scholar]
- 50.Egermann H, McAdams S. Empathy and emotional contagion as a link betwen recognised and felt emotions in music listening. Music Percept. 2013;31:139–156. [Google Scholar]
- 51.Koelsch S, Kilches S, Steinbeis N, Schelinski S. Effects of unexpected chords and of performer’s expression on brain responses and electrodermal activity. PLoS ONE. 2008;3:e2631. doi: 10.1371/journal.pone.0002631. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 52.Dellacherie D, Roy M, Hugueville L, Peretz I, Samson S. The effect of musical experience on emotional self-reports and psychophysiological responses to dissonance. Psychophysiology. 2011;48:337–349. doi: 10.1111/j.1469-8986.2010.01075.x. [DOI] [PubMed] [Google Scholar]
- 53.Steinbeis N, Koelsch S, Sloboda JA. Emotional processing of harmonic expectancy violations. Ann. N. Y. Acad. Sci. 2005;1060:457–461. doi: 10.1196/annals.1360.055. [DOI] [PubMed] [Google Scholar]
- 54.Sloboda J. Music structure and emotional response: Some empirical findings. Psychol. Music. 1991;19:110–120. [Google Scholar]
- 55.Blood AJ, Zatorre RJ. Intensely pleasurable responses to music correlate with activity in brain regions implicated in reward and emotion. PNAS. 2001;98:11818–11823. doi: 10.1073/pnas.191355898. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 56.Craig DG. An exploratory study of physiological changes during ‘chills’ induced by music. Music. Sci. 2005;9:273–287. [Google Scholar]
- 57.Grewe O, Kopiez R, Altenmüller E. The chill parameter: Goose bumps and shivers as promising measures in emotion research. Music Percept. 2009;27:61–74. [Google Scholar]
- 58.Guhn M, Hamm A, Zentner M. Physiological and musico-acoustic correlates of the chill response. Music Percept. 2007;24:473–484. [Google Scholar]
- 59.Rickard NS. Intense emotional responses to music: A test of the physiological arousal hypothesis. Psychol. Music. 2004;32:371–388. [Google Scholar]
- 60.Salimpoor VN, Benovoy M, Longo G, Cooperstock JR, Zatorre RJ. The rewarding aspects of music listening are related to degree of emotional arousal. PLoS ONE. 2009;4:e7487. doi: 10.1371/journal.pone.0007487. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 61.Grewe O, Nagel F, Kopiez R, Altenmüller E. Listening to music as a re-creative process: Physiological, psychological, and psychoacoustical correlates of chills and strong emotions. Music Percept. 2007;24:297–314. [Google Scholar]
- 62.Proverbio AM, et al. Non-expert listeners show decreased heart rate and increased blood pressure (fear bradycardia) in response to atonal music. Front. Psychol. 2015;6:1646. doi: 10.3389/fpsyg.2015.01646. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 63.Alluri V, et al. From Vivaldi to Beatles and back: Predicting lateralized brain responses to music. Neuroimage. 2013;83:627–636. doi: 10.1016/j.neuroimage.2013.06.064. [DOI] [PubMed] [Google Scholar]
- 64.Bernardi NF, et al. Increase in synchronization of autonomic rhythms between individuals when listening to music. Front. Physiol. 2017;8:1–10. doi: 10.3389/fphys.2017.00785. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 65.Bonetti, L. et al. Spatiotemporal brain dynamics during recognition of the music of Johann Sebastian Bach. bioRxiv 2020.06.23.165191 (2020). 10.1101/2020.06.23.165191.
- 66.Muszynski M, Kostoulas T, Lombardo P, Pun T, Chanel G. Aesthetic highlight detection in movies based on synchronization of spectators’ reactions. ACM Trans. Multimed. Comput. Commun. Appl. 2018;14:1–23. [Google Scholar]
- 67.Tschacher W, et al. Physiological synchrony in audiences of live concerts physiological synchrony in audiences of live concerts. Psychol. Aesthet. Creat. Arts. 2021 doi: 10.1037/aca0000431. [DOI] [Google Scholar]
- 68.Nastase SA, Gazzola V, Hasson U, Keysers C. Measuring shared responses across subjects using intersubject correlation. Soc. Cogn. Affect. Neurosci. 2019;14:669–687. doi: 10.1093/scan/nsz037. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 69.Hasson U, Nir Y, Levy I, Fuhrmann G, Malach R. Intersubject synchronization of cortical activity during natural vision. Science (80-) 2004;303:1634–1640. doi: 10.1126/science.1089506. [DOI] [PubMed] [Google Scholar]
- 70.Bacha-Trams M, Ryyppö E, Glerean E, Sams M, Jääskeläinen IP. Social perspective-taking shapes brain hemodynamic activity and eye movements during movie viewing. Soc. Cogn. Affect. Neurosci. 2020;15:175–191. doi: 10.1093/scan/nsaa033. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 71.Dmochowski JP, Sajda P, Dias J, Parra LC. Correlated components of ongoing EEG point to emotionally laden attention—A possible marker of engagement? Front. Hum. Neurosci. 2012;6:1–9. doi: 10.3389/fnhum.2012.00112. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 72.Golland Y, Arzouan Y, Levit-Binnun N. The mere co-presence: Synchronization of autonomic signals and emotional responses across co-present individuals not engaged in direct interaction. PLoS ONE. 2015;10:1–13. doi: 10.1371/journal.pone.0125804. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 73.Kauppi JP, Jääskeläinen IP, Sams M, Tohka J. Inter-subject correlation of brain hemodynamic responses during watching a movie: Localization in space and frequency. Front. Neuroinform. 2010;4:5. doi: 10.3389/fninf.2010.00005. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 74.Kang O, Wheatley T. Pupil dilation patterns spontaneously synchronize across individuals during shared attention. J. Exp. Psychol. Gen. 2017;146:569–576. doi: 10.1037/xge0000271. [DOI] [PubMed] [Google Scholar]
- 75.Schmälzle R, Häcker FEK, Honey CJ, Hasson U. Engaged listeners: Shared neural processing of powerful political speeches. Soc. Cogn. Affect. Neurosci. 2014;10:1137–1143. doi: 10.1093/scan/nsu168. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 76.Wilson SM, Molnar-Szakacs I, Iacoboni M. Beyond superior temporal cortex: Intersubject correlations in narrative speech comprehension. Cereb. Cortex. 2008;18:230–242. doi: 10.1093/cercor/bhm049. [DOI] [PubMed] [Google Scholar]
- 77.Ames DL, Honey CJ, Chow MA, Todorov A, Hasson U. Contextual alignment of cognitive and neural dynamics. J. Cogn. Neurosci. 2015;27:655–664. doi: 10.1162/jocn_a_00728. [DOI] [PubMed] [Google Scholar]
- 78.Simony E, et al. Dynamic reconfiguration of the default mode network during narrative comprehension. Nat. Commun. 2016;7:1–13. doi: 10.1038/ncomms12141. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 79.Herbec A, Kauppi JP, Jola C, Tohka J, Pollick FE. Differences in fMRI intersubject correlation while viewing unedited and edited videos of dance performance. Cortex. 2015;71:341–348. doi: 10.1016/j.cortex.2015.06.026. [DOI] [PubMed] [Google Scholar]
- 80.Kaneshiro B, Nguyen DT, Norcia AM, Dmochowski JP, Berger J. Natural music evokes correlated EEG responses reflecting temporal structure and beat. Neuroimage. 2020;214:116559. doi: 10.1016/j.neuroimage.2020.116559. [DOI] [PubMed] [Google Scholar]
- 81.Hasson U, Malach R, Heeger DJ. Reliability of cortical activity during natural stimulation. Trends Cogn. Sci. 2010;14:40–48. doi: 10.1016/j.tics.2009.10.011. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 82.Bracken BK, Alexander V, Zak PJ, Romero V, Barraza JA. Physiological synchronization is associated with narrative emotionality and subsequent behavioral response. In: Schmorrow DD, Fidopiastis CM, editors. Foundations of Augmented Cognition. Advancing Human Performance and Decision-Making through Adaptive Systems. Springer I; 2014. pp. 3–13. [Google Scholar]
- 83.Grey JM, Gordon JW. Perceptual effects of spectral modifications on musical timbres. J. Acoust. Soc. Am. 1978;63:1493. [Google Scholar]
- 84.Alluri V, Toiviainen P. Exploring perceptual and acoustical correlates of polyphonic timbre. Music Percept. 2012;27:223–242. [Google Scholar]
- 85.Wu, B., Horner, A. & Lee, C. Musical timbre and emotion: the identification of salient timbral features in sustained musical instrument tones equalized in attack time and spectral centroid. In Proceedings of 40th International Computer Music Conference. ICMC 2014 11th Sound Music Computer Conference. SMC 2014—Music Technology Meets Philosophy from Digital Echos to Virtual Ethos 928–934 (2014).
- 86.Krumhansl CL, Kessler EJ. Tracing the dynamic changes in perceived tonal organization in a spatial representation of musical keys. Psychol. Rev. 1982;89:334–368. [PubMed] [Google Scholar]
- 87.Krumhansl CL. Cognitive Foundations of Musical Pitch. Cognitive Foundations of Musical Pitch. Oxford University Press; 1990. [Google Scholar]
- 88.Toiviainen P, Krumhansl CL. Measuring and modeling real-time responses to music: The dynamics of tonality induction. Perception. 2003;32:741–766. doi: 10.1068/p3312. [DOI] [PubMed] [Google Scholar]
- 89.Hepokoski JA, Darcy W. Elements of Sonata Theory: Norms, Types, and Deformations in the Late-Eighteenth-Century Sonata. Oxford University Press; 2006. [Google Scholar]
- 90.Zbikowski L. M. Conceptualizing Music: Cognitive Structure, Theory, and Analysis. Oxford University Press; 2002. pp. 137–201. [Google Scholar]
- 91.Open Science collaboration. Estimating the reproducibility of psychological science. Science (80-).349 (2015). [DOI] [PubMed]
- 92.Cohen J. Statistical Power Analysis for the Behavioural Sciences. Academic Press; 1988. [Google Scholar]
- 93.Steinbeis N, Koelsch S, Sloboda JA. The role of harmonic expectancy violations in musical emotions: Evidence from subjective, physiological, and neural responses. J. Cogn. Neurosci. 2006;18:1380–1393. doi: 10.1162/jocn.2006.18.8.1380. [DOI] [PubMed] [Google Scholar]
- 94.Taruffi L, Pehrs C, Skouras S, Koelsch S. Effects of sad and happy music on mind-wandering and the default mode network. Sci. Rep. 2017;7:1–10. doi: 10.1038/s41598-017-14849-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 95.Ding N, et al. Temporal modulations in speech and music. Neurosci. Biobehav. Rev. 2017;81:181–187. doi: 10.1016/j.neubiorev.2017.02.011. [DOI] [PubMed] [Google Scholar]
- 96.Doelling KB, Arnal LH, Ghitza O, Poeppel D. Acoustic landmarks drive delta–theta oscillations to enable speech comprehension by facilitating perceptual parsing. Neuroimage. 2014;85(Pt 2):761–768. doi: 10.1016/j.neuroimage.2013.06.035. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 97.Gomez P, Danuser B. Affective and physiological responses to environmental noises and music. Int. J. Psychophysiol. 2004;53:91–103. doi: 10.1016/j.ijpsycho.2004.02.002. [DOI] [PubMed] [Google Scholar]
- 98.Fink LK, Hurley BK, Geng JJ, Janata P. A linear oscillator model predicts dynamic temporal attention and pupillary entrainment to rhythmic patterns. J. Eye Mov. Res. 2018;11:1–25. doi: 10.16910/jemr.11.2.12. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 99.Schroeder CE, Lakatos P. Low-frequency neuronal oscillations as instruments of sensory selection. Trends Neurosci. 2009;32:9–18. doi: 10.1016/j.tins.2008.09.012. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 100.Voloh B, Womelsdorf T. A role of phase-resetting in coordinating large scale neural networks during attention and goal-directed behavior. Front. Syst. Neurosci. 2016;10:1–19. doi: 10.3389/fnsys.2016.00018. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 101.Margulis EH. On Repeat: How Music Plays the Mind. Oxford University Press; 2014. [Google Scholar]
- 102.Cannam, C., Landone, C. & Sandler, M. Sonic visualiser: An open source application for viewing, analysing, and annotating music audio files. In MM’10 1467 (2010). 10.1145/1873951.1874248.
- 103.Tröndle M. Integrated methods: A pledge for integrative and interdisciplinary research. In: Vartanian O, Nadal M, editors. The Oxford Handbook of Empirical Aesthetics. Oxford University Press; 2020. [Google Scholar]
- 104.Merrill J, Czepiel A, Fink L, Toelle J, Wald-Fuhrmann M. The aesthetic experience of live concerts: Self-reports and psychophysiology. PsyArXiv. 2020 doi: 10.31234/osf.io/g829v. [DOI] [Google Scholar]
- 105.Müllensiefen D, Gingras B, Musil J, Stewart L. The musicality of non-musicians: An index for assessing musical sophistication in the general population. PLoS ONE. 2014;9:e89642. doi: 10.1371/journal.pone.0089642. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 106.Schaal NK, Bauer AKR, Müllensiefen DD. Gold-MSI: Replikation und Validierung eines Fragebogeninstrumentes zur Messung Musikalischer Erfahrenheit anhand einer deutschen Stichprobe. Music. Sci. 2014;18:423–447. [Google Scholar]
- 107.Vroegh TP. The Pleasures of Getting Involved into the Music: Absorption, and Its Role in the Aesthetic Appreciation of Music. Goethe-Universität Frankfurt; 2018. [Google Scholar]
- 108.Coutinho E, Scherer KR. Introducting the GEneva Music-induced Affect Checklist (GEMIAC): A brief instrument for the rapid assessment of musically induced emotions. Music Percept. 2017;34:1225–1228. [Google Scholar]
- 109.Schindler I, et al. Measuring aesthetic emotions: A review of the literature and a new assessment tool. PLoS ONE. 2017;12:8–11. doi: 10.1371/journal.pone.0178899. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 110.Lartillot, O. & Toiviainen, P. A matlab toolbox for musical feature extraction from audio. In Proceedings of 10th International Conference on Digital Audio Effects 237–244 (2007).
- 111.Tzanetakis G, Cook P. Musical genre classification of audio signals. IEEE Trans. Speech Audio Process. 2002;10:293–302. [Google Scholar]
- 112.Mencke I, Omigie D, Wald-Fuhrmann M, Brattico E. Atonal music: Can uncertainty lead to pleasure? Front. Neurosci. 2019;13:1–18. doi: 10.3389/fnins.2018.00979. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 113.Goodchild M, Wild J, McAdams S. Exploring emotional responses to orchestral gestures. Music. Sci. 2017;23:25–49. [Google Scholar]
- 114.Lange EB, Frieler K. Challenges and opportunities of predicting musical emotions with perceptual and automatized features. Music Percept. 2018;36:217–242. [Google Scholar]
- 115.Omigie D, et al. Experiencing musical beauty: Emotional subtypes and their physiological and musico-acoustic correlates. Psychol. Aesthet. Creat. Arts. 2019 doi: 10.1037/aca0000271. [DOI] [Google Scholar]
- 116.McAdams S, Winsberg S, Donnadieu S, De Soete G, Krimphoff J. Perceptual scaling of synthesized musical timbres: Common dimensions, specificities, and latent subject classes. Psychol. Res. 1995;58:177–192. doi: 10.1007/BF00419633. [DOI] [PubMed] [Google Scholar]
- 117.Alluri V, Toiviainen P. Exploring perceptual and acoustical correlates of polyphonic timbre. Music Percept. 2012;27:223–242. [Google Scholar]
- 118.Oostenveld R, Fries P, Maris E, Schoffelen JM. FieldTrip: Open source software for advanced analysis of MEG, EEG, and invasive electrophysiological data. Comput. Intell. Neurosci. 2011;2011:156869. doi: 10.1155/2011/156869. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 119.Piskorski J, Guzik P. Filtering Poincaré plots. Comput. Methods Sci. Technol. 2005;11:39–48. [Google Scholar]
- 120.Schoenberg A. In: Fundamentals of Musical Composition. Strang G, Stein L, editors. Faber and Faber; 1970. pp. 178–214. [Google Scholar]
- 121.R Development Core Team . R: A Language and Environment for Statistical Computing. R Development Core Team; 2017. [Google Scholar]
- 122.Revelle, W. psych: Procedures for Personality and Psychological Research, Northwestern University, Evanston, Illinois, USA. http://CRAN.R-project.org/package=psych, Version 1.6.4. (2016).
- 123.Bates, D., Mächler, M., Bolker, B. M. & Walker, S. C. Fitting linear mixed-effects models using lme4. J. Stat. Softw.67 (2015).
- 124.Kuznetsova A, Brockhoff PB, Christensen RHB. lmerTest package: Tests in linear mixed effects models. J. Stat. Softw. 2017;82:1–26. [Google Scholar]
- 125.Fox J, Weisberg S. Visualizing fit and lack of fit in complex regression models with predictor effect plots and partial residuals. J. Stat. Softw. 2018;87:1–27. [Google Scholar]
- 126.Lenth, R., Singmann, H., Love, J., Buerkner, P. & Herve, M. Package ‘emmeans’: Estimated marginal means, aka least squares means. R package version 1.4.8 (2020). Available at http://cran.r-project.org/package=qtl.
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
Data of this study are available from the corresponding author upon request.