Abstract
Neural, physiological, and behavioral signals synchronize between human subjects in a variety of settings. Multiple hypotheses have been proposed to explain this interpersonal synchrony, but there is no clarity under which conditions it arises, for which signals, or whether there is a common underlying mechanism. We hypothesized that cognitive processing of a shared stimulus is the source of synchrony between subjects, measured here as intersubject correlation (ISC). To test this, we presented informative videos to participants in an attentive and distracted condition and subsequently measured information recall. ISC was observed for electro-encephalography, gaze position, pupil size, and heart rate, but not respiration and head movements. The strength of correlation was co-modulated in the different signals, changed with attentional state, and predicted subsequent recall of information presented in the videos. There was robust within-subject coupling between brain, heart, and eyes, but not respiration or head movements. The results suggest that ISC is the result of effective cognitive processing, and thus emerges only for those signals that exhibit a robust brain–body connection. While physiological and behavioral fluctuations may be driven by multiple features of the stimulus, correlation with other individuals is co-modulated by the level of attentional engagement with the stimulus.
Keywords: intersubject correlation, brain–body connection, interpersonal physiological synchrony, engagement
Significance Statement.
Neural, physiological, and behavioral signals are synchronized between humans in a variety of settings. In this work, we show that physiological synchrony requires only 2 things, namely, effective cognitive processing of a common stimulus, and a robust coupling between brain activity and the physiological signal in question. We confirm this theory for heart rate, pupil size, gaze position, and saccade rate, as positive examples, and respiration and head movements as negative examples. We show that the strength of this correlation is co-modulated, i.e. correlation is modulated in unison for all signals in which correlation is detected. We propose that this common modulation is the result of attentional engagement with the stimulus.
Introduction
It is well established that shared experiences can synchronize physiological signals between individuals. This has been observed for autonomic signals (1) including galvanic skin response, heart rate (HR), body temperature, respiration, as well as other signals such as gaze position (2) and pupil size (3). Synchronization of physiological signals between individuals has often been attributed to physical or social interaction (4–7). However, the simultaneous experience is not a prerequisite for physiological synchronization, as a number of studies have shown that physiological signals can be correlated between subjects even when they engage individually with dynamic natural stimuli (3, 8–11).
Stimulus-induced correlation between individuals has been studied extensively with neural measures including functional magnetic resonance imaging (fMRI) (12), electro-encephalography (EEG) (13), magneto-encephalography (MEG) (14), and functional near-infrared spectroscopy (fNIRS) (15) in particular during presentation of film (2, 12), auditory narratives (16, 17), and music (18). These studies show that subjects process narrative stimuli similarly, and that correlation of brain activity is predictive of memory of the narrative (17, 19). The cognitive state of viewers has been shown to influence the intersubject correlation (ISC) of neural activity. For instance, subjects that are attentive (20) and more engaged have higher ISC (21). In total, these studies indicate that perceptual and cognitive processing of the narrative are similar across subjects and depend on cognitive state.
In the context of physiological or behavioral signals, ISC is often referred to as “interpersonal synchronization,” implying that 2 or more people are co-present in a given context. A variety of mechanisms have been proposed to cause ISC such as social interactions (1), physical interactions (4, 7, 22), shared emotions (9, 10), and it has been argued that the strength of synchrony is modulated by empathy (23, 24), arousal (4, 25), attention (26), and more (1). The diversity of factors parallels the diversity of factors known to affect physiological signals. For instance, HR fluctuations are often discussed in the context of emotions, and pupil size in the context of arousal, although both are affected by a variety of other cognitive factors (27–33). We hypothesize, instead, that the cognitive processing of a shared stimulus is sufficient to induce ISC, thus providing a more parsimonious explanation for the variety of phenomena previously observed. We use the term “cognitive processing” in its general sense of creation and manipulation of mental representations, which includes stimulus perception (34). The hypothesis predicts that ISC is co-modulated in different signals, and importantly, that it will emerge only for physiological signals that exhibit robust coupling to the brain. If, instead, physiological synchrony is driven by a variety of factors, we would not expect a co-modulation of ISC nor would we expect signals with robust brain–body coupling to necessarily synchronize between subjects.
To test these opposing predictions we collected physiological, neural, and behavioral signals while participants watched informative videos. Data was collected individually for each participant to rule out effects of direct social interactions. Additionally, videos were selected to be engaging but not to evoke strong emotions or arousal as in previous studies on HR synchronization (6, 8, 9). These controls can falsify alternative hypotheses that require social or physical interactions (1, 4, 7, 22) or strong emotions or empathy (8–10, 23, 24). Our hypothesis also predicts that ISC should be modulated by attention and predictive of memory of the content in the video. To test for this, we used a secondary mental task that distracts attention from the stimulus (20) and subsequently probed for recall memory (35). We measured ISC for each individual (12, 13), and resolved it on the group level also across time and frequency.
As predicted, we found significant ISC between individuals only for those signal modalities that exhibit a robust brain–body connection, namely, gaze position, pupil size, HR, and saccade rate. We did not find significant correlation for respiration or head velocity, which indeed did not exhibit a robust coupling with brain activity. Consistent with our hypothesis, the strength of ISC co-varied across signal modalities, was modulated by attention and was predictive of recall. These results suggest that ISC is modulated in unison by the level of attentional engagement with the stimulus.
Results
To establish the strength of ISC in different modalities we presented instructional videos while simultaneously recording neural, behavioral, and physiological signals. In the first experiment with 92 subjects (Experiment 1) we recorded the EEG, HR, gaze position (gaze), pupil size (pupil), and respiration. We chose 3 instructional videos related to physics, biology, and computer science, each 3–5 minutes long, with a total duration of 10 minutes. Signals were recorded at different sampling frequency for each modality but aligned in time across all subjects and modalities (Fig. 1A). For each signal modality we computed Pearson's correlation of these time courses between all pairs of subjects (Fig. 1B). For gaze position the correlation is computed separately for horizontal and vertical position and then averaged. For the 64 channels of EEG, we first extracted several components of the raw evoked potentials that maximally correlate between subjects (36), computed pairwise correlation between subjects for each component, and then took the sum of the correlation values. In each signal modality, ISC is defined for each subject as the average Pearson's correlation between that subject and all others (Fig 1C).
The first observation was that ISC is statistically significant in most subjects and modalities (black points in Fig. 1C), but also quite variable across individuals. Significant correlation was detected in all modalities except respiration. Comparing ISC across modalities, the most robust correlation was found for pupil size (ISCpupil: 0.06–0.51, mean M = 0.39, and SD = 0.09) and gaze position (ISCgaze is in the range of 0.01–0.37, M = 0.29, and SD = 0.06). They are significantly larger from chance values for all 92 subjects tested. Robust correlation is also found for EEG (ISCEEG: 0.00–0.11, M = 0.07, SD = 0.02, and 92/92 significant). We found that HR also correlates between subjects, but to a lesser extent with several subjects not exhibiting significant correlation (ISCHR: −0.01–0.10, M = 0.04, SD = 0.03, and 61/92 significant). For respiration we were not able to detect significant correlation between any participant and the group (ISCresp: −0.01–0.01, M = 0.00, and SD = 0.00).
ISC co-varies across different modalities
Given the strong variation of ISC across subjects, we wanted to determine if it co-varies in different signal modalities, i.e. if ISC is high in 1 modality, is it also high in other modalities? To this end we compared the ISC of different modalities across subjects (Fig. 2). Subjects with high ISC in EEG, HR, gaze position, and pupil size also showed high correlation in the other signal modalities. Correlation of ISC across subjects between these 4 modalities was significant in all cases (r = 0.49–0.69, P < 0.01 bonferroni corrected, and N = 92). However, we did not find any significant relation between the ISC of these 4 modalities and respiration (r = 0.04–0.25, P > 0.1 bonferroni corrected, and N = 92). The observation that the level of ISC of gaze position, pupil size and EEG co-varies across subjects is perhaps expected as they may all be driven by the visual dynamic of the video. What is more surprising is that high ISC in these modalities also coincided with high ISC of HR fluctuations, which thus far has been mostly attributed to emotional aspects of a stimulus 8 and would not be expected to be driven by visual dynamics of the video.
Time scale of correlated signal fluctuations and co-modulation across time
ISC captures whether subjects move their eyes in unison, whether their HR increases or decreases together, whether their pupils dilate or contract or whether they inhale and exhale simultaneously. Given that ISC is co-modulated between modalities, we wanted to know if these correlated fluctuations were due to similar entrainment with the stimulus. To investigate this, we resolved ISC by frequency band, i.e. the signals are band-pass filtered prior to computing ISC resulting in a coherence spectrum. These coherence spectra differed significantly between modalities (Fig. 3A). Therefore, co-modulation is not likely to result from entrainment to a specific frequency band, and instead may have been driven by more complex properties of the stimulus. While coherent fluctuations differed between modalities, they generally were slower than 10 Hz and are strong in the frequency band around 0.1 Hz. ISC might, therefore, be reliably measured on a time scale of 10 seconds. We additionally analyzed the power spectra, which quantifies the magnitude of fluctuations in different frequency bands (Fig. 3B). They differed significantly from the coherence spectrum. Therefore, coherence was frequency-specific and not just a result of the underlying dynamic. Given this diversity, it is possible that each signal modality was modulated by something different in the stimulus across time. We, therefore, asked whether ISC, from 1-time interval to the next, changes together in different modalities. Indeed, time-resolved ISC of brain signals computed on 10-second intervals correlate significantly with time-resolved ISC of gaze position, pupil size, and HR (with correlations in the range of r = −0.04–0.39), but not with respiration (Fig. 3C). This correlation of ISC between modalities over time was weaker than what we found across subjects. This suggests that on a short time scale (less than 15 minutes), ISC of different modalities may be driven by a diversity of factors in these video stimuli. Nevertheless, both across subjects and across time, we found that ISC of different signal modalities are co-modulated.
ISC is predictive of an individual's memory performance
We hypothesized that ISC is the result of cognitive processing, and thus we expected that the level of ISC will be predictive of memory performance. We, therefore, tested memory of the material in the video after presentation with a set of multiple choice questions (these data were available for 43 of the 92 subjects). Questions tested memory of the information presented in the video, which covered topics related to science, technology, and math (STEM). Questions probed for recognition and comprehension such as “What are stars made of?” or “How can Single Nuclear Polymorphism be useful?.” First, to quantify the common factor that is driving the co-modulation of ISC across subjects, we used principal component analysis (after z-scoring ISC values for each modality). The first principal component captured 52% of the population variance in ISC and the second component only 22% (Fig. 4A). The modalities captured by this first component were EEG, HR, gaze position, and pupil size (Fig. 4B), consistent with the co-variation of ISC observed in Fig. 2. Whereas the second principal component loaded mostly on respiration (Fig. 4B). As predicted, we found a strong correlation of memory performance with the first principal component of ISC (Fig. 4C) (r(39) = 0.59, P = 5.7⋅10−5, gaze position data was missing in 2 of the 43 subjects). In contrast, there was no significant correlation with the second component and test taking performance (r(39) = −0.23, and P = 0.14).
This result was confirmed when analyzing each modality independently (Fig. 4D and replicated for experiments 2 and 3 in Figure S5, Supplementary Material). ISC of EEG was predictive of students' test taking performance (r(41) = 0.57, and P = 7.1⋅10−5) as well as gaze position (r(39) = 0.60 and P = 3.2⋅10−5), and pupil size (r(39) = 0.51 and P = 6.7⋅10−4). To a lesser extent this was also true for HR (r(41) = 0.30 and P = 0.05). In contrast, ISC of respiration did not correlate with memory performance (r(41) = 0.01 and P = 0.94). We also analyzed the difference in ISC between men and women and did not find any significant difference in the modalities tested here (Figure S7, Supplementary Material).
Correlation with others depends on the individual's attention to the stimulus
The interpretation that cognitive processing of a common stimuli is required for ISC to emerge is consistent with the observation that it is modulated by attention for many of these modalities. Specifically, when subjects are distracted from the stimulus, ISC drops significantly for gaze position, pupil size, EEG as well as HR (3, 11, 20). To compare these effects across modalities and determine their time scales, we performed a new experiment in which N = 29 subjects watched videos while normally attending, and then again while distracted by a mental arithmetic task. In this Experiment 2, we used 6 instructional videos with a total duration of 31 minutes. We computed ISC again resolved by frequency but separately for the attentive and distracted conditions (Fig. 5A replicated for Experiment 3 in Figure S6, Supplementary Material). We found that ISC is significantly weaker in the distracted condition for all modalities (respiration was not measured in this experiment). We also analyzed the power spectrum of these signals to determine if the attentional effects are reflected in the dynamic of these signals (Fig. 5B). We found generally weaker effect sizes for the power spectrum (Fig. 5C). This suggests that attention does not strongly affect physiological dynamics, but rather, fluctuations align in time across subjects when attending to the stimulus and not otherwise. Finally, we analyzed traditional measures of arousal involving pupil size and HR (Figure S4, Supplementary Material). Only HR variability (HRV) was meaningfully modulated by attention.
ISC occurs only for physiological signals that are coherent with brain signals
Given our hypothesis, we predicted that significant ISC in the physiological signals occurs if, and only if the signal correlates with brain activity within subjects during video watching. In this view, when brains correlate, so will other physiological signals that are driven by cognitive processing in the brain. The alternative hypothesis is that the brain–body connection itself is modulated by attention for these modalities. To test for this, we measured the within-subjects coupling between EEG and HR, pupil size, and respiration (Fig. 6). Specifically, at each frequency band we extracted a component of the EEG that best correlated with the respective signals. Here, we report the strength of this correlation, which we call within-subject correlation (WSC). We found that gaze position, pupil size, and HR significantly correlated with brain signals in some frequency bands (Fig. 6, and Figure S3 (Supplementary Material) for distribution over the scalp). This is trivially expected for gaze position due to saccade-evoked potentials (37), but this has not been previously reported for pupil or HR. Importantly, we found no correlation between brain potentials and respiration. We found only minor differences between the attend and distract conditions in these coherence spectra (for respiration we only have data on the attentive condition). The picture that emerges, therefore, is that the brain correlates with other subjects when it engages with the stimulus, and that this carries over to physiological signals as a result of an endogenous brain–body coupling, which is relatively stable with regards to attentional state during passive video watching.
Brain–body connection predicts ISC for novel signals
Our theory, therefore, is that signals correlate between subjects, but only if there is a robust coupling between brain signals and the physiological or behavioral signal in question (Fig. 7A). We tested this on 2 additional signal modalities that have independently been proposed as markers of arousal. One is head velocity (38), following the basic notion of arousal as movement of the body. The other is saccade rate, which has been recently linked to effort and task engagement (39). First, we measured whether either are coupled to the EEG (on data from Experiment 2), and found this to be the case for saccade rate but not head velocity. We, therefore, predicted that saccade rate and not head velocity will be correlated across subjects and this correlation will be modulated by attention and predict memory. These predictions were all confirmed by subsequent analysis (Fig. 7B and 7C), and indeed, consistent with this theory, we found a co-modulation of ISC between all modalities across subjects (Figure S1, Supplementary Material) and across time (Figure S2, Supplementary Material). Note that correlated gaze position does not necessarily imply correlated saccade rate, as it only captures the number of saccade per unit time. On the other hand, head velocity could have been expected to correlate as large saccades tend to be accompanied by corresponding reorienting of the head (40). Therefore, these positive and negative examples were not trivially anticipated without the proposed theory.
Discussion
To summarize, during viewing of informative videos we found significant ISC in all physiological signals that exhibited robust coupling with the EEG. This was the case for HR, pupil size, gaze position, and saccade rate, but not respiration or head velocity. This is consistent with the theory that correlation between individuals is the result of cognitive processing of a shared stimulus, but will only be observed for those physiological or behavioral signals that exhibit robust coupling with brain activity (Fig. 8). Importantly, the strength of ISC was co-modulated across signals, was predictive of an individual's memory performance and was attenuated when subjects were distracted from the stimulus.
The strength of ISC observed here as well as the modulation with attention and correlation with memory performances are in line with previous reports for HR (11), gaze position, pupil size (3), and EEG (17, 20, 35). The novel finding is that ISC is co-modulated in these signals across subjects and across time. We provide a novel theory as to which signals do or do not correlate and confirm this with saccade rate and head velocity as positive and negative examples.
The co-modulation observed here is surprising given the diversity of factors that have been previously proposed to underlie interpersonal physiological synchrony. Leaving out brain signals, this includes emotions (HR (5–7, 10)), arousal (HR (4, 25) and head velocity (38)), empathy (HR (6, 8, 24), skin conductance (23)), attention (eye movement (3) and HR (25, 26)), auditory features (HR and respiration (41, 42)), visual features (gaze position (43, 44)), physical movement (respiration (45, 46) and HR (22)), social interaction (HR (47)), and more. We argue more generally, that cognitive processing of the stimulus is a minimum requirement to evoke any of these constructs in a viewer's mind, and indeed, it suffices to explain the attention and memory effect we observed, namely, without processing the information imparted in the stimulus it is not possible to recall this information, and distracting the viewer will disrupt processing and thus reduce ISC. If interpersonal synchronization was really the result of the various factors as postulated in prior literature, it is not clear how this diversity of factors could have affected different modalities similarly, and even less that they should have modulated ISC in unison across time and across subjects. That said, the weak co-variation observed across time suggests that a diversity of factors may drive signal fluctuations differentially in each modality. Note that studies on interpersonal synchrony are often conducted with subject interacting, or at least co-present in the same space. These factors may have additional synchronizing effects (5, 7). What we showed here is that this is not a prerequisite for physiological synchrony.
An alternative interpretation of the present results is that low-level stimulus features are the source of co-variation in ISC. For example, visual dynamics may attract similar eye movement and this affects ISC of brain signals. With the strength of visual dynamics itself varying over time this could cause a co-modulation of ISC of the eyes and brain. There may be any number of stimulus features that guide visual exploration and could cause such co-variation in a bottom-up manner. However, this alternative interpretation does not explain the co-modulation of ISC observed across subjects. It also does not readily explain the co-modulation in heart and eyes, nor the observation that ISC of the heart is predictive of memory, a phenomenon that persists for audio-only narratives (11). In contrast, the effects of attention and on memory observed here are straightforward consequences of the cognitive processing hypothesis, which in our definition includes perceptual processing of the stimulus.
So, what might be the common factor that affects cognitive processing of the stimulus, and co-modulates the strength of ISC? We speculate that it is related to engagement with the stimulus. In previous work, with EEG during passive exposure to narrative audio and video we introduced the concept of “attentional engagement” with the stimulus (13, 17, 35, 48). We demonstrated that an objective behavioral measure of stimulus engagement correlates with ISC of the EEG measured on time scales of 10 seconds (49), presumably reflecting fluctuations of attention (20). In the vision literature, “attentional engagement” implies not only attracting attention (gaze) but also processing features of the target of attention (50–53). The notion that the ISC of EEG reflects a level of engagement with the stimulus has been adopted by a number of groups (21, 54, 55). The new finding here is that this engagement may be a common factor that also modulates the strength of fluctuations in HR, pupil size, saccade rate, and gaze position.
Regardless of the interpretation, this co-modulation of physiological responses is surprising for another reason—the dynamics of each of these signal fluctuations and their underlying neural control are quite diverse. For instance, neural control of HR and pupil size are ascribed to different brain structures. While pupil dilation has been most closely linked to activity in superior colliculus (SC) and locus coeruleus (LC) (29), HR is controlled by midbrain structures that are modulated by input from the amygdala, cingulate, and insula (27), although the LC also projects to nuclei involved in cardiac regulation (56). Additionally, the dynamics of these signals, as we have seen in their power spectra, differ considerably between modalities and the attentional effect on ISC manifests in different frequency bands for each. We note that the coupling between EEG and each of the signal modalities did not differ significantly with attentional state. Note also that the effects of attention on the power spectrum were generally small. Therefore, cognitive processing does most likely not substantially change the ongoing endogenous physiological dynamic within subjects. Instead, it appears to only time-align the existing dynamic to the external stimulus.
Our theory led us to prospectively analyse the correlation between EEG and various physiological signals. While a correlation of EEG with gaze position is well-established (in fact, it is considered a nuisance artefact (57)), to our knowledge this is the first study to report a correlation of EEG with HR, pupil size, and saccade rate. The theory also correctly predicted that saccade rate would correlate across subjects, something that has not been reported previously in the literature. Even less obvious was that this correlation of saccade rate would be modulated by attention and predict subsequent memory performance.
Here, we have focused on first-order correlation as we had no specific hypotheses related to nonlinear interactions. Such interactions can give rise to cross-frequency coupling in phase or amplitude and may warrant more complex analyses (58, 59).
We did not find ISC in breathing, which one might have expected. HR naturally increases when we inhale and decreases with exhale. This is known as respiratory sinus arrhythmia (60). Since HR synchronizes, we might have expected that breathing also does. Alternatively, in moments of suspense, viewers might be “holding their breath” in unison, which might have resulted in synchronized breathing. Breathing seems to align to the timing of a behavioral task (61), but here participants were not asked to do anything physical during viewing. Generally, studies that show correlated breathing involve some concurrent physical activity such as singing (46), speaking (62), or dancing (45) and correlation is not observed during passive listening to speech (63), though some effects have been found when listening to music (41). We do not report results on breathing rate, as it was not expected to fluctuate much during rest and would have required longer recordings. But we do not rule out that breathing rate may synchronize.
The fact that gaze positions are robustly correlated between subjects might not come as a surprise since eye movements are heavily driven by the dynamic of visual stimuli (64). We deliberately chose stimuli with compelling visual dynamics, such as animated graphics. A caveat of our analysis is that eye movements contaminate EEG signals, and thus ISC of eye movements could trivially result in ISC of EEG. However, we have previously shown that this confound can be largely ruled out when removing the EOG artefacts as we have done here (20). Furthermore, the co-modulation of ISCEEG with other modalities is present even during audio-only narratives and in the absence of eye movements (in preparation). Another well-known artefact in the EEG is due to the heart beat itself. However, note that heart beat does not actually synchronize between subjects and so this artefact is not likely to contribute to the ISC of the EEG. Instead, the synchrony we detected between subjects is in the fluctuations of their HR. To our knowledge, the link we report here between EEG and HR has not been reported previously in the literature. One additional caveat is that at low frequencies (below 0.1Hz) the EEG may also include galvanic skin response from the scalp, however we still find ISC of EEG and WSC for signals above this frequency.
The pupil size is known to be affected by luminance, but also higher level factors such as problem solving and cognitive effort (31, 65), affective processing (33, 66), and attention (67). So, to see the pupil size correlated across subjects might be expected. That this pupil correlation should be modulated by attentional engagement is consistent with the argument that task engagement is reflected in pupil dynamics on similarly short time scales (39). Our most recent work suggests that this phenomenon is not contingent on luminance fluctuations as it appears also during audio-only narratives (in preparation). That the ISC of pupil should be co-modulated with ISC of HR was not expected from the literature. We have not analyzed eye blinks, however, given reports of a correlation of blinking across subjects during movies (68), we expect that this will behave similarly to the other measures we have taken from the eyes. Given the coupling that has been observed between EEG and gastric rhythms (69) we also predict that these will show significant ISC.
Pupil and HR have often been linked to arousal (see below), although the term “arousal” has quite diverging definitions in the literature. We expected that head velocity could serve as a marker of physical arousal (38), and thus may also correlate. Correlation may also be expected given the ISC of gaze position and the fact that large eye movements tend to be accompanied by corresponding head movements (40). We did not find significant ISC in head velocity in our smaller datasets. Consistent with that, ISC of head velocity did not significantly correlate with memory or attention, nor did head velocity significantly correlate with brain potentials. We did, however, find a weak ISC in the larger dataset of 92 subjects (see Supplement), and therefore, one might be able to resolve these effects on larger datasets. Indeed, our theory does not stipulate an all-or-nothing effect. A weak brain–body link may lead to similarly weak physiological ISC and related effects.
Correlation of HR between subjects has previously been linked to emotional processing of video stimuli (8, 9), which may not have been a main driving factor in our mostly informative videos. In social contexts, HR synchronization has also been linked to empathy (70) and social bond (71). We propose that cognitive processing of the natural stimulus is a requirement to infer the emotional valence of the film (9), for an audience to bond with a performer (4) or mother and child to interpret social cues (47). Thus, cognitive processing of the external natural stimulus is the common denominator, even for interpersonal HR synchrony. A caveat to this conclusion is that we did not look for other factors that could have driven the different modalities differently. All videos were designed to communicate factual information related to science and technology, and not, for instance, evoke suspense or strong emotions (5, 6, 10, 72). So it is possible that we did not find other factors simply because we did not manipulate other factors with our experimental paradigm. Future work should consider varying stimulus properties along different dimensions.
The cognitive control of HR is a crucial prerequisite for the co-modulation observed here. Indeed, central control is well-established and explains the effects of cognition on HR and HR variability (27, 28). For instance, HRV correlates with neural activity measured with fMRI (73). There is also evidence for the opposite causal direction, whereby volitional control of breathing affects HR and in turn neural activity (74). The heart beat itself seems to affect saccade timing (75), evoke EEG (76), as well as visual detection performance (77). Also, strong stressors can increase HR and at the same time enhance the magnitude of evoked responses in EEG (78). But none of these studies report a link between EEG and HR as we have found here, nor do they anticipate that ISC is co-modulated in these modalities
A recent theory proposes a common drive to pupils and saccades (39). The theory posits that this common drive is affected by a variety of cognitive processes, in particular “task engagement“, but authors distinguish this from ”arousal". This common drive is postulated to be the activity in SC. Indeed, the SC along with the LC are midbrain nuclei that mediate cognitive effects on the pupil (29, 79). A common drive of pupil and saccades reconciles the dependence of pupil dilations on arousal and effort as well as its fluctuations during a task. Consistent with this we find that pupil size and saccade rate both behave similarly with attention and memory and importantly the ISC in these 2 modalities are co-modulated. Note that the LC also connects to midbrain structures that control cardiac function (80) and may mediate cognitive effects on HR (56). Therefore, the LC is well-situated to mediate cognitive effects on pupil size, saccades rate, and HR.
In conclusion, this work brings together 2 separate research fields which have demonstrated correlation between subjects in ecologically valid scenarios, namely, interpersonal physiological synchrony (1) and ISC of neural activity (12). We postulate that cognitive processing of a natural dynamic stimulus is required to drive and coordinate behavioral responses of the body. Our rationale is that humans have evolved to quickly respond to the natural stimuli in their environment, and to do so, the brain processes information and prepares the body to react. This leads to rapid, but reliable fluctuations of HR, pupil size, gaze position, and others on the scale of seconds. If the brain is not processing the stimulus effectively, i.e. similarly across subjects, then the body will not respond appropriately and these fluctuations will no longer be guided by the stimulus. As the brain changes in attentional state on time scales of 10 seconds or longer the correlation induced across subjects is modulated for different modalities in unison. Although our study here is only correlational, we do postulate a causal effect of cognitive processing on physiological and behavioral responses that is similar across subjects and modalities. Evidently, there could be a bidirectional interaction. This is most obvious for eye movements, as our view can affect our cognitive state. A bottom-up effect has also been hypothesized for heart signals (74) and is the basis for some meditation practices that focus on breathing (81). Establishing a causal direction of these effects will be difficult but is a worthwhile topic for future research.
Materials and methods
Participants
A total of 158 subjects participated in 1 of 3 experiments, where they watched informative videos. In Experiment 1, N = 96 participated (51 Female, age 18–49, M = 25.33, and SD = 7.29; 4 subjects were removed due to bad timing or bad signal quality). In Experiment 2, N = 32 subjects participated (21 females, age 18–57, M = 25.93, and SD 8.94 years; 3 subjects were removed due to bad signal quality). Lastly in Experiment 3, N = 31 participated (19 females, age 18–50, M = 25.79, and SD = 8.13 years; 2 subjects were removed due to bad signal quality).
Stimuli
A complete list of the video stimuli is provided in Table S1 (Supplementary Material). These stimuli have been used in previous studies (3, 11, 35). For Experiment 1, we selected 3 videos from YouTube channels that post short informal instructional videos, namely “Kurzgesagt—In a Nutshell” and “Minute Physics.” The videos covered topics related to physics and biology with a short duration (Range: 3–6.5 minutes and total duration 15.5 minutes). In Experiment 2, we selected 6 videos from “Khan Academy,” “eHow,” “Its ok to be smart,” and “SciShow,” which are popular online educational channels on YouTube. The videos covered topics related to biology and physics, with a short duration (Range: 4.5–6.5 minutes and total duration 31 minutes). In Experiment 3, we selected 5 informal instructional videos again from YouTube, covering topics related to physics, biology, and computer science with a short duration (Range: 2.4–6.5 minutes and Average: 4.1 +/− 2.0 minutes). Data on gaze position and pupil size for Experiment 2 have been previously analyzed (3), as well as data on HR from Experiment 3 (11). All other data and analyses are new to this study, i.e. all of the data from Experiment 1, EEG, HR, saccade rate, and head velocity from Experiment 2, and EEG, pupil size, gaze position, saccade rate, and head velocity from Experiment 3.
Procedure
All experiments were carried out at the City College of New York with approval from the Institutional Review Boards of the City University of New York. Documented informed consent was obtained from all subjects at the start of the experiment. Subject watched the videos on a 19” monitor while seated comfortably in a sound-attenuated booth with white fabric walls and normal ambient LED lighting placed around the subject. Sound was played through stereo speakers placed at 60° angles from the subject next to the monitor, both at a distance of approximately 60 cm from the subject.
In Experiment 1, subjects watched 3 instructional videos while their EEG, electro-oculogram (EOG), electro-cardiogram (ECG), respiration, pupillary responses, gaze, and head position were recorded. In the second experiment, subjects watched 6 instructional videos while their EEG, ECG, pupillary responses, gaze, and head position were recorded.
In all the experiments, subjects were instructed to watch the videos normally as they would at home, while being relaxed and sitting still. We refer to this as the attentive conditions (A). After they had watched the videos, subjects were given a 4-alternative forced-choice questionnaire covering factual information imparted during the video (11–12 recognition questions per video; see Table S1, Supplementary Material). The videos and question pairs were presented in random order. In Experiment 1, subjects were not aware that they would be tested on the material, whereas in Experiments 2 and 3 the test was anticipated. After answering questions, in Experiments 2 and 3 subjects were asked to watch the videos again, but this time to silently count in their mind backwards in steps of 7 (starting from a prime number picked at random between 800 and 1,000). The second viewing with concurrent counting is referred to as the distracted condition (D).
For segmentation of the physiological signals we used common onset and offset triggers, in addition, a flash and beep sound was embedded right before and after each video, which were recorded using a StimTracker (Cedrus) to ensure precise alignment across all subjects. To enable all modalities to be on the same time scale, we used the lab streaming layer (LSL) protocol. In addition, triggers were sent to both the eye tracking and EXG recording systems, timestamps from each system were used in a linear regression model to convert timestamp between each modality estimated using the common triggers.
Recording and preprocessing of EEG
The EEG was recorded at a sampling frequency of 2,048 Hz using a BioSemi Active Two system. Participants were fitted with a standard, 64-electrode cap following the international 10/10 system with the ground electrode located next to POz. In addition, the EOG was recorded with 6 auxiliary electrodes (1 located dorsally, ventrally, and laterally to each eye). The EEG and EOG data is band-pass filtered between 0.016 and 250 Hz by the Active two system prior to sampling. The signal was then digitally high-pass filtered (0.05Hz cut-off) and notch filtered at 60 Hz to remove line noise. To remove artefacts and outliers Robust PCA was used (82), and subsequently the signal was low-pass filtered (64 Hz cut-off) and down-sampled to 128 Hz. Bad electrode channels were identified manually and replaced with interpolated channels. The interpolation was performed using the 3D Cartesian coordinates from the electrode cap projected onto a plane using all surrounding “good” electrodes. The EOG channels were used to remove eye-movement artefacts by linearly regressing them from the EEG channels, i.e. least-squares noise cancellation (code for robust PCA and EOG noise cancelling in EEG can be found in https://www.parralab.org/isc/). In each EEG channel, additional outlier samples were identified as values exceeding 4 times the distance between the 25th and the 75th quartile of the median-centred signal, and samples 40 ms before and after such outliers were replaced with interpolated samples using neighboring electrodes.
Recording and preprocessing of ECG
The ECG signal was recorded using 2 ECG electrodes placed below the left collar bone and 1 on the left lumbar region with a BioSemi Active Two system at a sampling frequency of 2,048 Hz. The ECG signal was detrended using a high-pass filter (0.5 Hz cut-off) and subsequently notch filtered at 60 Hz to remove line noise. Peaks in the ECG corresponding to the R-waves were found using findpeaks (built-in Matlab function). The instantaneous HR is computed for each beat as the inverse of time intervals between subsequent R-wave peaks as in (11). To ensure the same sampling frequency for all subjects this instantaneous HR signal is resampled at a regular sampling rate of 128 Hz.
Recording and preprocessing of gaze position, head velocity, and pupil size
Gaze position, head movements and pupil size were recorded using the Eyelink 1000 eye tracker (SR Research Ltd. Ottawa, Canada) with a 35 mm lens at a sampling frequency of 500 Hz. Subjects were instructed to sit still while the experiment was carried out, but were free to move their heads, to ensure comfort (no chin rest). A standard 9-point calibration scheme was used utilizing manual verification. Stable pupillary responses were ensured by adjusting the background color of the calibration screen and all instructions presented to the subjects to be the average luminance of all the videos presented during the experiment. After each stimulus presentation, a drift-check was performed and the eye tracker was recalibrated if the visual angular error was greater than 2°. Blinks were detected using the algorithm of the eye tracker. These artefacts, blinks and 100 ms before and after were filled with linearly interpolated values. Head velocity is computed as the absolute value of the analytic signal of the Hilbert transform (root mean square summed over the 3 directions). Saccades were detected by the algorithm of the eye tracker. Instantaneous saccade rate was calculated as the inverse time interval between saccades and up-sampled to a constant sampling rate of 2,000 Hz to match the other eye tracking signals.
Recording and preprocessing of respiration
The respiration signal was recorded at a sampling frequency of 2,048 Hz on the BioSemi Active Two system using a SleepSense 1387 Respiratory Effort Sensors, which captures the tension on a belt worn around the chest of the subject. The polarity of the signal was detected using peaks in the respiration signal and inverted to ensure the correct phase of the signal.
ISC analysis of gaze position, pupil size, respiration, HR, saccade rate, and head velocity
For each of the gaze position, pupil size, respiration, HR, saccade rate and head velocity signals, the ISC was computed in the following 3 steps: (1) computing the Pearson's correlation coefficient between a single subject's signal (each of the 6 modalities independently) and that of all other subjects while they watched a video. Correlation is computed by summing over all time points of the video (with durations specified in Table S1, Supplementary Material) and then averages over all videos. (2) A single ISC value for a subject was obtained by averaging the correlation values between that subject and all other subjects. (3) The 2 first steps are then repeated for all subjects, resulting in a single ISC value for each subject. For ISC of gaze position, we compute the ISC in the horizontal and vertical gaze direction using the procedure as described above separately. To obtain 1 single ISCgaze value, we average the ISC for the horizontal and vertical directions.
ISC of EEG
For the EEG signals ISC was computed using correlated component analysis (36), with code available at http://parralab.org/corrca/. This method finds linear components of the EEG that are most correlated between subjects. The components consist of several projection vectors that linearly combine electrodes, on which the data is projected. The ISC of each component is obtained by computing the correlation coefficients of the projected EEG between each participant and all other participants. We only use components that are significantly correlated above chance (circular shuffle on test set data, see below). This yielded 3–9 components depending which of the 3 Experiments was analyzed. ISC values are then summed over all significant components.
Statistical significance of ISC values per subject
To determine whether ISC values are significantly larger than 0 (Figs 1C and 7D), we determine the null distribution of ISC values on surrogate data obtained with circular shuffle statistics (83). P-value (type I error rate) is then the fraction of shuffles with ISC values larger than in the original unshuffled data. We performed 10,000 circular shuffles estimating P-values down to 0.0001. For EEG, the ISC values are measured on components that have been optimized to provide high ISC values. To avoid an upwards bias in ISC during statistical testing, components are optimized on training data and significance is established for separate test data (2 video clips are used for training and 1 for testing). Note that all statistical comparisons of correlation values, here and in the remainder, are performed by computing correlation values with the identical number of samples for the conditions to be compared.
Frequency analysis of ISC (coherence spectrum)
We performed a frequency analysis to investigate at which time scale the recorded signals correlated between subjects. Each signal from each subject was band-pass filtered using 5th order Butterworth filters with logarithmic spaced center frequencies with a bandwidth of 0.2 of the center frequency. The ISC was computed for each subject in each frequency band for all videos. To obtain a single ISC value per frequency band we average ISC values for all videos and subjects.
Computation of d-prime (d’) effect of attention on coherence and power spectrum
To determine the effect size of the attentional modulation for both the frequency-resolved ISC and Power spectrum we compute the d-prime statistics. For each frequency band (see section above) we compute the ISC and power for each subject in the attending and distracted conditions. We quantify the effect size between the 2 conditions as the difference of the means divided by the standard deviation across subjects of the differences (paired Cohen's d).
Cluster statistics for difference between attentive and distracted conditions
To determine significant difference in spectra between attending and distracted conditions (for Figs 5A, 5B, 6B, 7A, and 7B), we use cluster shuffle statistics as follows. Since different frequency bands in the frequency-resolved ISC (and WSC) are not independent, we use 1D cluster statistics including random shuffles to correct for multiple comparisons following an established procedure (84, 85). Briefly, this procedure involves 4 steps: (1) take the difference between the spectrum in the attending and the distracted condition for each subject. (2) compute a one-sample t test on this difference for each frequency band. (3) clusters are identified as consecutive frequency bands with P-values below 0.01. The t-values within each cluster are then summed. (4) Run 10,000 permutations in which we randomly change on half of the subject's the sign of difference between the spectra computed in step 1. Steps 2 and 3 are then repeated while keeping the sum of t-values of the largest cluster. Finally, we compare the clusters’ t-values obtained in step 3 with the distribution of permuted cluster t-values obtained in step 4. Clusters with larger than 99% (corresponding to P-value < 0.01) of the permuted distribution were considered significant after multiple comparison cluster correction. Note that for the contrast in attention all data was used for optimization (of components of the EEG) including attentive and distracted conditions, so that any difference cannot be due to the optimization procedure.
Cluster statistics for significant of intersubject coherence spectra
To determine if frequency-resolved ISC (and WSC) values are significantly different from 0 (Figs 3A, 3B, and 6A), we use a similar cluster statistic as above. For cases involving EEG (correlated component analysis and regression), we use test data to avoid upwards bias due to optimization, which is performed on separate training data. The shuffling and cluster correction procedure consists of the same 4 steps as above, except that we divide subjects in 2 equal size groups at random. The premise of this is that values around 0 will not differ significantly if placed at random in 2 different groups (84).
WSC analysis (brain–body coupling)
To determine the coupling between the brain and different signal modalities we compute WSC using ordinary least squares regression, i.e. we measure how well each signal modality can be predicted linearly from the multidimensional EEG signal. Specifically, denote with the EEG signal at time t for in channel i, and with the signal of the modality of interest. The linear prediction is then and we find weights such that is minimized. Linear projections of this sort are conventionally referred to as components of the EEG (as in principal component analysis, or correlated component analysis used for the ISC). WSC is then the Pearson's correlation of and extending over all stimuli. In the case where we differentiate between the attend and distract conditions, we compute component weights and WSC in each condition separately. For the frequency-resolved WSC, we band-pass filter the signal modality and EEG in the same fashion as the time-resolved ISC, namely using 5th order Butterworth filters with logarithmic spaced center frequencies with a bandwidth of 0.2 of the center frequency. To compute the WSC in the frequency resolved condition, we find components for each band separately. WSC in all instances was computed on test data using leave-one-subject out cross-validation (i.e. regression parameters are estimated including all subjects except for the test subject; WSC is then computed on that test subject; this train-test process is then repeated for all subjects.) Statistical significance is established in the same way as for ISC as described in the 2 previous sections.
Common factor analysis of ISC
To find the common factor of ISC between signal modalities, we first remove outliers of the ISC data that are larger or smaller than 4 times the interquartile difference of the data. This was done since the standardization of the data is sensitive to outliers. We then standardize the ISC (0 mean and unit variance) and compute the principal components.
Supplementary Material
Acknowledgments
We would like to thank Christopher J. Honey and Michael P. Milham for discussions on early versions of this work.
Notes
Competing Interest: The authors declare no competing interest.
Contributor Information
Jens Madsen, Department of Biomedical Engineering, City College of New York, 160 Convent Avenue, New York, NY 10031, USA.
Lucas C Parra, Department of Biomedical Engineering, City College of New York, 160 Convent Avenue, New York, NY 10031, USA.
Funding
This work was funded by the National Science Foundation through grant number DRL-1660548.
Authors' Contributions
J.M and L.C.P designed the experiments. J.M collected the data. J.M and L.C.P analyzed the data. L.C.P and J.M wrote the manuscript.
References
- 1. Palumbo RV, et al. . 2017. Interpersonal autonomic physiology: a systematic review of the literature. Pers Soc Psychol Rev. 21:99–141. [DOI] [PubMed] [Google Scholar]
- 2. Hasson U, et al. . 2008. Neurocinematics: the neuroscience of film. Projections. 2:1–26. [Google Scholar]
- 3. Madsen J, Júlio SU, Gucik PJ, Steinberg R, Parra LC. 2021. Synchronized eye movements predict test scores in online video education. Proc Natl Acad Sci.118:e2016980118. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4. Konvalinka I, et al. . 2011. Synchronized arousal between performers and related spectators in a fire-walking ritual. Proc Natl Acad Sci. 108:8514–8519. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5. Golland Y, Arzouan Y, Levit-Binnun N. 2015. The mere co-presence: synchronization of autonomic signals and emotional responses across co-present individuals not engaged in direct interaction. PLoS ONE. 10:e0125804. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6. Kaltwasser L, et al. . 2019. Sharing the filmic experience – the physiology of socio-emotional processes in the cinema. PLoS ONE. 14:e0223259. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7. Ardizzi M, Calbi M, Tavaglione S, Umiltà MA, Gallese V. 2020. Audience spontaneous entrainment during the collective enjoyment of live performances: physiological and behavioral measurements. Sci Rep. 10:1–12. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8. Steiger BK, Kegel LC, Spirig E, Jokeit H. 2019. Dynamics and diversity of heart rate responses to a disaster motion picture. Int J Psychophysiol. 143:64–79. [DOI] [PubMed] [Google Scholar]
- 9. Golland Y, Keissar K, Levit-Binnun N. 2014. Studying the dynamics of autonomic activity during emotional experience: dynamics of autonomic activity in emotional experience. Psychophysiology. 51:1101–1111. [DOI] [PubMed] [Google Scholar]
- 10. Bracken BK, Alexander V, Zak PJ, Romero V, Barraza JA. 2014. Physiological synchronization is associated with narrative emotionality and subsequent behavioral response. In: Schmorrow DD, Fidopiastis CM, editors. Foundations of augmented cognition. Advancing human performance and decision-making through adaptive systems. Springer International Publishing, Crete. p. 3–13. [Google Scholar]
- 11. Pérez P, et al. . 2021. Conscious processing of narrative stimuli synchronizes heart rate between individuals. Cell Rep. 36:109692. [DOI] [PubMed] [Google Scholar]
- 12. Hasson U, Nir Y, Levy I, Fuhrmann G, Malach R. 2004. Intersubject synchronization of cortical activity during natural vision. Science. 303:1634–1640. [DOI] [PubMed] [Google Scholar]
- 13. Dmochowski JP, et al. . 2014. Audience preferences are predicted by temporal reliability of neural processing. Nat Commun. 5:4567. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14. Lankinen K, Saari J, Hari R, Koskinen M. 2014. Intersubject consistency of cortical MEG signals during movie viewing. Neuroimage. 92:217–224. [DOI] [PubMed] [Google Scholar]
- 15. Pan Y, et al. . 2020. Instructor-learner brain coupling discriminates between instructional approaches and predicts learning. Neuroimage. 211:116657. [DOI] [PubMed] [Google Scholar]
- 16. Simony E, et al. . 2016. Dynamic reconfiguration of the default mode network during narrative comprehension. Nat Commun. 7:12141. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17. Cohen SS, Parra LC. 2016. Memorable audiovisual narratives synchronize sensory and supramodal neural responses. eNeuro. 3:ENEURO.0203–ENEU16.2016. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18. Madsen J, Margulis EH, Simchy-Gross R, Parra LC. 2019. Music synchronizes brainwaves across listeners with strong effects of repetition, familiarity and training. Sci Rep. 9:3576. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19. Hasson U, Furman O, Clark D, Dudai Y, Davachi L. 2008. Enhanced intersubject correlations during movie viewing correlate with successful episodic encoding. Neuron. 57:452–462. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20. Ki JJ, Kelly SP, Parra LC. 2016. Attention strongly modulates reliability of neural responses to naturalistic narrative stimuli. J Neurosci. 36:3092–3101. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21. Schmälzle R, Häcker FEK, Honey CJ, Hasson U. 2015. Engaged listeners: shared neural processing of powerful political speeches. Soc Cognit Affect Neurosci. 10:1137–1143. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22. Gordon I, et al. . 2020. Physiological and behavioral synchrony predict group cohesion and performance. Sci Rep. 10:8484. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23. Marci CD, Ham J, Moran E, Orr SP. 2007. Physiologic correlates of perceived therapist empathy and social-emotional process during psychotherapy. J Nerv Mental Dis. 195:103–111. [DOI] [PubMed] [Google Scholar]
- 24. Goldstein P, Weissman-Fogel I, Shamay-Tsoory SG. 2017. The role of touch in regulating inter-partner physiological coupling during empathy for pain. Sci Rep. 7:3252. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25. Mitkidis P, McGraw JJ, Roepstorff A, Wallot S. 2015. Building trust: heart rate synchrony and arousal during joint action increased by public goods game. Physiol Behav. 149:101–106. [DOI] [PubMed] [Google Scholar]
- 26. Stuldreher IV, Thammasan N, van Erp JBF, Brouwer A-M. 2020. Physiological synchrony in EEG, electrodermal activity and heart rate reflects shared selective auditory attention. J Neural Eng. 17:046028. [DOI] [PubMed] [Google Scholar]
- 27. Thayer JF, Hansen AL, Saus-Rose E, Johnsen BH. 2009. Heart rate variability, prefrontal neural function, and cognitive performance: the neurovisceral integration perspective on self-regulation, adaptation, and health. Ann Behav Med. 37:141–153. [DOI] [PubMed] [Google Scholar]
- 28. Thayer JF, Lane RD. 2009. Claude Bernard and the heart–brain connection: further elaboration of a model of neurovisceral integration. Neurosci Biobehav Rev. 33:81–88. [DOI] [PubMed] [Google Scholar]
- 29. Joshi S, Gold JI. 2020. Pupil size as a window on neural substrates of cognition. Trends Cogn Sci. 24:466–480. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30. van der Wel P, van Steenbergen H. 2018. Pupil dilation as an index of effort in cognitive control tasks: A review. Psychon Bull Rev. 25:2005–2015. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 31. Hess EH, Polt JM. 1964. Pupil size in relation to mental activity during simple problem-solving. Science. 143:1190–1192. [DOI] [PubMed] [Google Scholar]
- 32. Preuschoff K, t'Hart BM, Einhauser W. 2011. Pupil dilation signals Surprise: evidence for Noradrenaline's role in decision making. Front Neurosci. 5:115. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 33. Bradley MM, Miccoli L, Escrig MA, Lang PJ. 2008. The pupil as a measure of emotional arousal and autonomic activation. Psychophysiology. 45:602–607. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 34. Krch D. 2011. Cognitive processing. In: Kreutzer JS, DeLuca J, Caplan B, editors. Encyclopedia of clinical neuropsychology. Berlin, Heidelberg: Springer. p. 627–627.. DOI: 10.1007/978-0-387-79948-3_1443. [Google Scholar]
- 35. Cohen SS, et al. . 2018. Neural engagement with online educational videos predicts learning performance for individual students. Neurobiol Learn Mem. 155:60–64. [DOI] [PubMed] [Google Scholar]
- 36. Parra LC, Haufe S, Dmochowski JP. 2019. Correlated components analysis – extracting reliable dimensions in multivariate data. Neurons Behav Data Anal Theory. 2:7125. [Google Scholar]
- 37. Parra LC, Spence CD, Gerson AD, Sajda P. 2005. Recipes for the linear analysis of EEG. Neuroimage. 28:326–341. [DOI] [PubMed] [Google Scholar]
- 38. Wass SV, de Barbaro K, Clackson K. 2015. Tonic and phasic co-variation of peripheral arousal indices in infants. Biol Psychol. 111:26–39. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 39. Burlingham CS, Mirbagheri S, Heeger DJ. 2021. A unified model of the task-evoked pupil response. bioRxiv. DOI: 10.1101/2021.04.09.439231 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 40. Stahl JS. 1999. Amplitude of human head movements associated with horizontal saccades. Exp Brain Res. 126:41–54. [DOI] [PubMed] [Google Scholar]
- 41. Bernardi NF, et al. . 2017. Increase in synchronization of autonomic rhythms between individuals when listening to music. Front Physiol. 8: 785. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 42. Garssen B. 1979. Synchronization of respiration. Biol Psychol. 8:311–315. [DOI] [PubMed] [Google Scholar]
- 43. Franchak JM, Heeger DJ, Hasson U, Adolph KE. 2016. Free viewing gaze behavior in infants and adults. Infancy. 21:262–287. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 44. Wang HX, Freeman J, Merriam EP, Hasson U, Heeger DJ. 2012. Temporal eye movement strategies during naturalistic viewing. J Vis. 12:16–16. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 45. Codrons E, Bernardi NF, Vandoni M, Bernardi L. 2014. Spontaneous group synchronization of movements and respiratory rhythms. PLoS ONE. 9:e107538. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 46. Müller V, Lindenberger U. 2011. Cardiac and respiratory patterns synchronize between persons during choir singing. PLoS ONE. 6:e24893. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 47. Feldman R, Magori-Cohen R, Galili G, Singer M, Louzoun Y. 2011. Mother and infant coordinate heart rhythms through episodes of interaction synchrony. Infant Behav Dev. 34:569–577. [DOI] [PubMed] [Google Scholar]
- 48. Dmochowski JP, Sajda P, Dias J, Parra LC. 2012. Correlated components of ongoing EEG point to emotionally laden attention – a possible marker of engagement?. Front Hum Neurosci. 6:112. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 49. Cohen SS, Henin S, Parra LC. 2017. Engaging narratives evoke similar neural activity and lead to similar time perception. Sci Rep. 7:4578. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 50. Posner MI, Walker JA, Friedrich FJ, Rafal RD. 1984. Effects of parietal injury on covert orienting of attention. J Neurosci. 4:1863–1874. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 51. Zivony A, Lamy D. 2018. Contingent attentional engagement: stimulus- and goal-driven capture have qualitatively different consequences. Psychol Sci. 29:1930–1941. [DOI] [PubMed] [Google Scholar]
- 52. Folk CL, Ester EF, Troemel K. 2009. How to keep attention from straying: Get engaged!. Psychon Bull Rev. 16:127–132. [DOI] [PubMed] [Google Scholar]
- 53. Adam KCS, Mance I, Fukuda K, Vogel EK. 2015. The contribution of attentional lapses to individual differences in visual working memory capacity. J Cogn Neurosci. 27:1601–1616. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 54. Bevilacqua D, et al. . 2018. Brain-to-brain synchrony and learning outcomes vary by student–teacher dynamics: evidence from a real-world classroom electroencephalography study. J Cogn Neurosci. 31:401–411. [DOI] [PubMed] [Google Scholar]
- 55. Dikker S, et al. . 2017. Brain-to-brain synchrony tracks real-world dynamic group interactions in the classroom. Curr Biol. 27:1375–1380. [DOI] [PubMed] [Google Scholar]
- 56. Samuels ER, Szabadi E. 2008. Functional neuroanatomy of the noradrenergic locus coeruleus: its roles in the regulation of arousal and autonomic function part I: principles of functional organisation. Curr Neuropharmacol. 6:235–253. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 57. Hillyard SA, Galambos R. 1970. Eye movement artifact in the CNV. Electroencephalogr Clin Neurophysiol. 28:173–182. [DOI] [PubMed] [Google Scholar]
- 58. Rassi E, Dorffner G, Gruber W, Schabus M, Klimesch W. 2019. Coupling and decoupling between brain and body oscillations. Neurosci Lett. 711:134401. [DOI] [PubMed] [Google Scholar]
- 59. Marwan N, Carmen Romano M, Thiel M, Kurths J. 2007. Recurrence plots for the analysis of complex systems. Phys Rep. 438:237–329. [Google Scholar]
- 60. Hirsch JA, Bishop B. 1981. Respiratory sinus arrhythmia in humans: how breathing pattern modulates heart rate. Am J Physiol Heart Circulat Physiol. 241:H620–H629. [DOI] [PubMed] [Google Scholar]
- 61. Johannknecht M, Kayser C. 2021. The influence of the respiratory cycle on reaction times in sensory-cognitive paradigms. Sci Rep. 16: 2586. DOI: 10.1101/2021.05.06.442800. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 62. Rochet-Capellan A, Fuchs S. 2014. Take a breath and take the turn: how breathing meets turns in spontaneous dialogue. Philos Trans R Soc B Biol Sci. 369:20130399. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 63. Rochet-Capellan A, Fuchs S. 2013. Changes in breathing while listening to read speech: the effect of reader and speech mode. Front Psychol. 4: 906. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 64. Dorr M, Martinetz T, Gegenfurtner KR, Barth E. 2010. Variability of eye movements when viewing dynamic natural scenes. J Vis. 10:28–28. [DOI] [PubMed] [Google Scholar]
- 65. Piquado T, Isaacowitz D, Wingfield A. 2010. Pupillometry as a measure of cognitive effort in younger and older adults. Psychophysiology. 47:560–569. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 66. Partala T, Surakka V. 2003. Pupil size variation as an indication of affective processing. Int J Hum Comput Stud. 59:185–198. [Google Scholar]
- 67. Mathôt S, Van der Stigchel S. 2015. New light on the mind's eye: the pupillary light response as active vision. Curr Dir Psychol Sci. 24:374–378. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 68. Nakano T, Yamamoto Y, Kitajo K, Takahashi T, Kitazawa S. 2009. Synchronization of spontaneous eyeblinks while viewing video stories. Proc R Soc B Biol Sci. 276:3635–3644. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 69. Richter CG, Babo-Rebelo M, Schwartz D, Tallon-Baudry C. 2017. Phase-amplitude coupling at the organism level: the amplitude of spontaneous alpha rhythm fluctuations varies with the phase of the infra-slow gastric basal rhythm. Neuroimage. 146:951–958. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 70. Järvelä S, Kivikangas JM, Kätsyri J, Ravaja N. 2014. Physiological linkage of dyadic gaming experience. Simul Gam. 45:24–40. [Google Scholar]
- 71. Feldman R. 2012. Parent–infant synchrony: a biobehavioral model of mutual influences in the formation of affiliative bonds. Monogr Soc Res Child Dev. 77:42–51. [Google Scholar]
- 72. Song H, Finn ES, Rosenberg MD. 2020. Neural signatures of attentional engagement during narratives and its consequences for event memory. Proc Natl Acad Sci USA. 17:e2021905118. DOI: 10.1101/2020.08.26.266320. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 73. Thayer JF, Åhs F, Fredrikson M, Sollers JJ, Wager TD. 2012. A meta-analysis of heart rate variability and neuroimaging studies: Implications for heart rate variability as a marker of stress and health. Neurosci Biobehav Rev. 36:747–756. [DOI] [PubMed] [Google Scholar]
- 74. Mather M, Thayer JF. 2018. How heart rate variability affects emotion regulation brain networks. Curr Opin Behav Sci. 19:98–104. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 75. Galvez-Pol A, McConnell R, Kilner JM. 2020. Active sampling in visual search is coupled to the cardiac cycle. Cognition. 196:104149. [DOI] [PubMed] [Google Scholar]
- 76. Luft CDB, Bhattacharya J. 2015. Aroused with heart: modulation of heartbeat evoked potential by arousal induction and its oscillatory correlates. Sci Rep. 5:15717. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 77. Park H-D, Correia S, Ducorps A, Tallon-Baudry C. 2014. Spontaneous fluctuations in neural responses to heartbeats predict visual detection. Nat Neurosci. 17:612–618. [DOI] [PubMed] [Google Scholar]
- 78. Eason RG, Dudley LM. 1970. Physiological and behavioral indicants of activation. Psychophysiology. 7:223–232. [DOI] [PubMed] [Google Scholar]
- 79. Wang C-A, Munoz DP. 2015. A circuit for pupil orienting responses: implications for cognitive modulation of pupil size. Curr Opin Neurobiol. 33:134–140. [DOI] [PubMed] [Google Scholar]
- 80. Aston-Jones G, et al. . 1991. Chapter 4 – afferent regulation of locus coeruleus neurons: anatomy, physiology and pharmacology. In: Barnes CD, Pompeiano O, editors. Progress in brain research. Amsterdam: Elsevier. vol. 88, p. 47–75. [DOI] [PubMed] [Google Scholar]
- 81. Zeidan F, et al. . 2011. Brain mechanisms supporting the modulation of pain by mindfulness meditation. J Neurosci. 31:5540–5548. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 82. Lin Z, Chen M, Ma Y. 2013. The augmented lagrange multiplier method for exact recovery of corrupted low-rank matrices. J Struct Biol. 181:116–127.23110852 [Google Scholar]
- 83. Parra LC, Haufe S, Dmochowski JP. 2019. Correlated components analysis – extracting reliable dimensions in multivariate data. Neurons, Behavior, Data analysis, and Theory. 2(1): [Google Scholar]
- 84. Cohen MX. 2014. Analyzing neural time series data: theory and practice. Cambridge (MA): MIT Press. [Google Scholar]
- 85. Maris E, Oostenveld R. 2007. Nonparametric statistical testing of EEG- and MEG-data. J Neurosci Methods. 164:177–190. [DOI] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.