Abstract
Complex auditory sequences known as music have often been described as hierarchically structured. This permits the existence of non-local dependencies, which relate elements of a sequence beyond their temporal sequential order. Previous studies in music have reported differential activity in the inferior frontal gyrus (IFG) when comparing regular and irregular chord-transitions based on theories in Western tonal harmony. However, it is unclear if the observed activity reflects the interpretation of hierarchical structure as the effects are confounded by local irregularity. Using functional magnetic resonance imaging (fMRI), we found that violations to non-local dependencies in nested sequences of three-tone musical motifs in musicians elicited increased activity in the right IFG. This is in contrast to similar studies in language which typically report the left IFG in processing grammatical syntax. Effects of increasing auditory working demands are moreover reflected by distributed activity in frontal and parietal regions. Our study therefore demonstrates the role of the right IFG in processing non-local dependencies in music, and suggests that hierarchical processing in different cognitive domains relies on similar mechanisms that are subserved by domain-selective neuronal subpopulations.
Introduction
Complex auditory sequences known as music exist in all human cultures1, and elements in many musical styles are hierarchically structured2,3. Examples include harmonic progressions in classical western tonal music4 and jazz5, as well as transformations of tone rows in twelve-tone serialist compositions6. A sequence is said to be hierarchical if the dependencies (rules which bind two elements) between its elements can be represented as a type of mathematical graph called a rooted tree (acyclic graph with a designated root element)7,8. This means that all elements of a sequence are connected to form an overarching structure7, and implies the existence of a sub-/superordinate relationship between elements9. We shall refer to this definition of hierarchy throughout the text. Furthermore, a dependency is said to be local if it relates elements that directly follow one another in a sequence, and non-local if the dependency spans over multiple intervening elements. Consider nested sequences, which have the form AnAn−1…A1B1…Bn−1Bn and contain dependencies that are embedded, or nested, within another dependency: In the nested sequence A3A2A1B1B2B3, the (local) dependency between A1 and B1 is embedded within the (non-local) dependency between A2 and B2, which is in turn embedded within the (non-local) dependency between A3 and B3. Importantly, the ability to relate remote, non-local musical events beyond their immediate temporal sequential order is said to be crucial for successfully processing hierarchical structures in music7,10,11.
Previous studies on processing hierarchical structures in music argued that humans can differentiate between auditory tone sequences generated according to a hierarchical recursive rule and an iterative rule12, show priming effects in integrating harmonic contextual information13,14, and discriminate between grammatical and ungrammatical transformations in serialist music15,16. Moreover, harmonically irregular chords within a chord sequence were shown to elicit an early right anterior negativity (ERAN) in event-related brain potentials (ERPs) (using EEG11,17,18 and MEG for the magnetic equivalent19), which can already be observed in infants20, and in a musical scale previously unheard by participants21. Functional MRI (fMRI) studies employing similar violation paradigms have also reported differential responses in the bilateral inferior frontal gyri (sometimes with a right-hemispheric dominance)18,22–27, and the anterior insular cortices22,23,27. Activity in the inferior frontal gyrus (IFG) was suggested to process hierarchical structure in music, given that language studies have implicated the left IFG – particularly the pars opercularis – in syntactic reordering and embedding28–31. This is corroborated by interference effects observed between musical and linguistic syntax32–34, which suggest that processing hierarchical structure relies on domain-general neural resources35.
However, the question of whether hierarchical processing was really involved in the previous studies has also been raised. It has been argued that humans rearrange scrambled phrases of music in a way that was grammatically-coherent locally but not globally36, and are insensitive towards transpositions37,38 and re-orderings39 to sections of classical music pieces. Moreover, violations to the hierarchical structure in previous experiments were not restricted to dependencies between non-local elements, but also violated local dependencies between immediately-adjacent chords11. Although the ERP study by Koelsch and colleagues11 controlled for the possibility of processing musical sequences in a strictly local fashion, the precise neuro-functional basis of processing non-local dependencies in music nevertheless remains unknown.
The current study was thus conducted to assess the functional basis of processing non-local dependencies in music in the human brain, whilst controlling for local transition probabilities. Our approach was a grammaticality judgment task based on an artificial grammar learning paradigm40–42. Musicians learnt a nested atonal grammar of piano-tone sequences (i.e. AnAn−1…A1B1…Bn−1Bn) before discriminating between novel grammatical and ungrammatical musical sequences during fMRI scanning. In contrast to previous studies, successful completion of the task requires participants to explicitly abstract notes into motifs and to store multiple non-local dependencies in parallel. We moreover manipulated the level of embedding (that is, the number of nested dependencies in a sequence) to dissociate the processing of nested dependencies in music from the effects of increasing working memory demands on processing these dependencies, and to ensure that the observed responses in resolving the nested dependencies generalised to different levels.
We hypothesised that violations to the nested grammar would elicit increased BOLD responses in the bilateral inferior frontal gyri, especially in the right hemisphere. Based on the literature on auditory tonal working memory43–46, we also hypothesised increased BOLD responses in the dorsolateral prefrontal cortex and parietal areas with increased levels of embedding due to additional working memory demands.
Materials and Methods
Participants
Twenty musically-trained participants (12 females, 8 males) with normal hearing and at least seven years of training (cf.47–49) in their most experienced instrument (M = 13.30 years, SD = 5.78 years) completed both sessions of the experiment. No participants reported absolute pitch or neurological/psychological disorders. Participants were excluded from further analyses if their hit rates during the fMRI session for grammatical sequences, ungrammatical sequences with category violations, or ungrammatical sequences with state violations (see Stimuli) fell below the 5% significance level of performing above chance according to a binomial test. Two male participants were excluded that way. One female participant was excluded due to incidental findings. Data were analysed for the remaining 17 participants (mean age = 26.29 years, SD = 2.37 years; mean experience in most experienced instrument = 13.76 years, SD = 6.00 years, seven of whom were conservatory-level). They were right-handed50, and had a mean score of 93.3 (SD = 13.1) in general music sophistication51. Informed consent was obtained from all participants and the experiment was approved by the ethics committee of the University of Leipzig in accordance with the Declaration of Helsinki.
Stimuli
Our study employed a 2 × 2 factorial design with factors GRAMMATICALITY (GRAMMATICAL versus UNGRAMMATICAL) and LEVEL OF EMBEDDING (LoE: ONE-LoE versus TWO-LoE) to dissociate the effects of grammaticality and working memory in processing non-local dependencies in music.
Auditory sequences of the nested structure AnAn−1…A1B1…Bn−1Bn (see Fig. 1A) were generated by concatenating two (ZERO-LoE; structure A1B1), four (ONE-LoE; structure A2A1B1B2), or six musical motifs (TWO-LoE; structure A3A2A1B1B2B3). Each motif consisted of three successive isochronous piano tones (duration = 250 ms per tone) and belonged to one of four categories in one of two states (see Fig. 1A, Table 1 and Supplementary Audio clips SA1–6). The motifs of a sequence were randomly concatenated without replacement so that distinct categories in state A preceded the same categories in state B but in reverse order. This prevented participants from using a counting strategy. Ungrammatical sequences that violated this nested structure were introduced by interchanging either the state (so-called state violations) or category (so-called category violations), but not both, of exactly one B-motif. The replaced motif in a category violation was chosen so that the category was not presented before in the sequence. All sequences contained a pause (duration = 750 ms) between each pair of motifs, and the lowest note of each motif in a sequence was uniformly sampled between all twelve tones in the octave between C4 and B4. This was to avoid participants relying solely on tonal information and accomplishing the task based on matching tones they heard. Sequence durations were thus 2.25 s for ZERO-LoE, 5.25 s for ONE-LoE, and 8.25 s for TWO-LoE.
Table 1.
Motif category | State | |
---|---|---|
A | B | |
Arpeggio | 0-4-7 | 7-4-0 |
Scale | 0-2-3 | 3-2-0 |
Jump | 0-0-7 | 7-0-0 |
Arrow | 0-9-0 | 9-0-9 |
Note: The lowest pitch of each motif was between C4 and B4, and was unique for each motif in a nested sequence.
Due to the extensive number of unique sequences (see Supplementary Information), ONE-LoE and TWO-LoE sequences were presented only once throughout the entire experiment. Sequences with ZERO-LoE were used without replacement before being reshuffled back into the pool after each session.
Procedure
The experiment was divided into a training session and a scanning session, which were around 3.5 weeks apart (mean = 25.76 days, SD = 7.20 days); only participants who acquired the musical grammar in the training session participated in the scanning session (76.67% of participants successfully acquired the grammar). The experiment was programmed on Presentation 18.1 (Neurobehavioral Systems, Inc., Berkeley, CA, USA). Auditory stimuli (mono, 44,100 Hz sampling rate, 16 bits per sample) were delivered at a comfortable volume through circumaural headphones (with foam earplugs inserted inside the MRI scanner). White text was shown against a black background on a computer screen or viewed using a mirror attached to the head-coil from a back-projected image. Foam pads were placed to reduce head movement before fMRI scanning.
In the training session, participants were instructed to learn the grammar of a new language and extract the rule underlying the sequences. The session resembled the learning phase by Bahlmann and colleagues40 (see Supplementary Information).
In the scanning session (see Fig. 1B), participants discriminated inside the MRI scanner the grammaticality of 144 novel sequences, equally divided between the four combinations of the two factors: GRAMMATICALITY (GRAMMATICAL versus UNGRAMMATICAL) and LEVEL OF EMBEDDING (ONE-LoE versus TWO-LoE). Violations in ungrammatical sequences were counterbalanced for violation type (i.e. state versus category) and occurrence amongst the B positions. Stimuli were presented across six runs with a break (25 s) between each run. Participants were notified visually 5 s before the end of each break. Stimuli were pseudo-randomised such that at most two consecutive stimuli shared the same LEVEL OF EMBEDDING and GRAMMATICALITY. Each trial began with a randomly-jittered fixation cross (1.2 s–2.2 s) at the centre of the screen, followed by the presentation of an auditory sequence. After a short break (0.75 s), the letters Y (grammatical) and N (ungrammatical) appeared on the lower left and right sides of the fixation cross, for which participants had a 4 s time-window to decide on the grammaticality of the preceding sequence by pressing either the right index or middle finger on an MR-compatible button box; The letter position was pseudo-randomised. Visual feedback was displayed on the centre of the screen (1 s) to motivate performance (see Data analysis) and the next trial ensued. Additionally, 18 grammatical and 18 ungrammatical 0-LoE sequences were presented as filler sequences and were not analysed.
Data acquisition
Imaging data were collected on a 3 T Magnetom Skyra scanner (Siemens Healthcare, Erlangen, Germany) with a 20-channel head coil. Slices were acquired axially parallel to the AC-PC line for the whole-brain using a gradient EPI sequence (31 slices per volume, slice thickness = 3 mm, inter-slice gap = 1 mm, acquisition order = odd-interleaved ascending, FoV = 192 mm × 192 mm, acquisition matrix = 64 × 64, TR = 2000 ms, TE = 30 ms, flip angle = 90°, bandwidth = 2004 Hz/Px, echo spacing = 0.56 ms). The functional scan time was 42.3 minutes and 1270 volumes were obtained continuously. T1-weighted structural images (voxel size = 1 mm isotropic) of each participant were used to coregister and normalise the functional images to MNI space.
Data analysis
Data were analysed using MATLAB R2016a (The MathWorks, Inc., Natick, MA, USA), JASP 0.7.5.6 (JASP Team), and R 3.3.3 (The R Foundation for Statistical Computing, Vienna, Austria). To measure how well participants discriminated between GRAMMATICAL and UNGRAMMATICAL nested musical sequences, we applied signal detection theory52 to dissociate their behavioural sensitivity in detecting deviants in grammaticality from their response bias using the non-parametric sensitivity measure A and associated log-bias ln(b)53. The sensitivity measure gives an estimate of the mean area under the ROC curve, and the dissociation of sensitivity and bias avoids a misrepresentation of performance due to conflated hit rates. A high sensitivity score thus corresponds to a high hit rate and a low false positive rate. As we wanted participants to perform accurately, they were not instructed to respond as quickly as possible and reaction times were not analysed.
After checking for normality using the Shapiro-Wilk test, one-sample t-tests were conducted to compare the overall mean sensitivity against chance-level (0.5), and mean bias against 0, as well as paired t-tests to compare mean sensitivity and bias for the two LEVELS OF EMBEDDING separately. Effect sizes were calculated using Cohen’s d for correlated samples54.
Imaging data were analysed using SPM 12.6685 (Wellcome Trust Centre for Neuroimaging, London, UK). Volumes were slice-timing corrected, realigned to the first volume using rigid-body transformation and motion-susceptibility correction, coregistered to the individual’s structural image, resampled to a voxel size of 3 mm × 3 mm × 4 mm and normalised to MNI space, and smoothed with a FWHM Gaussian kernel 2.5 times the voxel size for preprocessing.
For statistical analyses at the first-level, a voxelwise GLM was estimated for each participant. Each sequence was modelled as a boxcar function of the same stimulus duration and convolved with the canonical HRF. One regressor was used to model correctly-responded sequences for each combination of the two factors: GRAMMATICALITY (GRAMMATICAL versus UNGRAMMATICAL) and LEVEL OF EMBEDDING (ONE-LoE versus TWO-LoE). A regressor modelling all remaining sequences, a regressor indicating volumes within breaks, a regressor to remove finger-press artifacts, and six motion regressors were added as regressors of no interest. Finger presses were modelled using a boxcar function with the response-prompt as onset and reaction time as duration and were convolved with the canonical HRF. A high-pass filter (128 s cut-off) and an autoregressive AR(1) model were applied.
To guarantee that the observed effects on grammaticality were not confounded by serial processing, we additionally estimated a refined model at the subject-level where ungrammatical sequences only consisted of correctly-responded category violations at positions B2 and B3 post hoc. Sequences with violations in position B1 or state violations could have been rejected by a strategy which does not require resolving non-adjacent dependencies, as it might have been possible for participants to have detected these violations by only comparing surface features of the violated motif and the preceding motif55,56. However, it is unlikely that participants relied solely on serial or local processing to reject these sequences, because (1) they could not have known beforehand the grammaticality of a sequence and position of the violation, and so would have had built a nested representation of the motifs to complete the task in case the sequence was grammatical, and (2) only participants who were proficient in detecting category violations were included in the analysis.
For statistical analyses at the group-level, data were modelled by a 2 × 2 flexible factorial model. We assumed independence at the subject and condition levels, equal variance in the former and unequal in the latter. Significant clusters were identified using an a priori-defined voxelwise FWE-corrected threshold of p < 0.05 and an extent of four voxels, determined by rounding above the expected cluster size estimated from the smoothness of the SPM based on Gaussian random-field theory57. Anatomical locations were identified using the MNI2TAL tool58 and SPM Anatomy Toolbox 2.2c59.
We turned to a psychophysiological interaction (PPI) analysis to explore the extent to which brain regions implicated in the current study are functionally related. In brief, PPI measures the change in functional connectivity between two regions under different experimental contexts60. This allows us to infer any (undirected) flow of task-relevant information between two brain regions. Given our factorial design, we carried out a generalised PPI analysis61 as opposed to a traditional PPI analysis62. Generalised PPI is more suited to our experiment as the model includes all psychological factors and hence spans the entire experimental space.
Seed regions were defined by drawing spheres (radii = 4.5 mm, chosen to avoid overlap between the regions of interest) around the maxima of each significant cluster of the refined model. They were then multiplied with the group mask to ensure that each seed region only included brain voxels present in all participants. We examined the task-modulated functional connectivity between each seed region and the remaining clusters across the four conditions by comparing the gPPI regressors of each subject in a 2 × 2 flexible factorial model at the group-level. An explicit mask of the seed regions was applied during the estimation of the second-level model, and we adopted the same statistical threshold as before. The analysis was performed using the gPPI toolbox 13.161.
To further examine whether psychophysiological interactions predicted behavioural performance, we additionally correlated the mean difference in beta estimates of each significant PPI cluster with participants’ overall sensitivity in detecting grammatical violations. The beta estimates of each cluster were extracted using MarsBaR 0.4463.
Data analysed during this study are available from the corresponding author upon reasonable request.
Results
Behavioural results
Participants’ overall sensitivity (M = 0.95 (out of a maximum score of 1), SD = 0.04, range = 0.88–1.00) in discriminating between grammatical and ungrammatical nested musical sequences was significantly above chance (t(16) = 51.92, p = 1.44 × 10−19, Cohen’s d = 12.59) and correlated with the number of years of training in their most experienced instrument (r = 0.63, p(corrected) = 0.03, see Figure S1 and Supplementary Information), although the response bias (M = −0.19, SD = 0.17) indicated a significant bias towards judging a sequence as grammatical (t(16) = −4.52, p = 3.47 × 10−4, Cohen’s d = −1.10). The mean sensitivity for ONE-level of embedding (LoE) sequences was also significantly higher than TWO-LoE sequences, but the difference in bias between the two levels of embedding was not significant (see Fig. 2 and Table 2).
Table 2.
LEVEL OF EMBEDDING | t(16) | p-value | Cohen’s d | ||
---|---|---|---|---|---|
ONE | TWO | ||||
Sensitivity | 0.96 (0.04) | 0.94 (0.04) | 2.73 | 0.02* | 0.66 |
Bias | −0.16 (0.23) | −0.22 (0.17) | 1.22 | 0.24 | 0.30 |
Notes: Values inside brackets indicate standard deviation. Maximum possible sensitivity score is 1; A negative bias indicates a tendency to judge a sequence as grammatical; *p < 0.05.
Imaging results
Distinct clusters of significant BOLD-response differences (see Fig. 3A and Table 3) were evaluated using SPM t-contrasts at the whole-brain level for both main effects of GRAMMATICALITY and LEVEL OF EMBEDDING. For the main effect of GRAMMATICALITY (UNGRAMMATICAL > GRAMMATICAL), we found a cluster of increased BOLD response with maxima in the right inferior frontal gyrus (IFG; pars opercularis, triangularis, and orbitalis), right middle frontal gyrus (in the dorsolateral prefrontal cortex), and right anterior insular cortex (AIC). We additionally identified clusters in the pre-supplementary motor area (pre-SMA), right dorsal premotor cortex, and left anterior insular cortex (AIC). The reverse contrast (GRAMMATICAL > UNGRAMMATICAL) yielded increased responses in the bilateral ventromedial prefrontal cortex (vmPFC).
Table 3.
Anatomical region | BA | Cluster size (voxels) | MNI-coordinates | t-value | ||
---|---|---|---|---|---|---|
X | Y | Z | ||||
UNGRAMMATICAL > GRAMMATICAL | ||||||
R anterior insular cortex (AIC) | 663 | 33 | 23 | −2 | 8.42 | |
R pars opercularis | 44 | 51 | 17 | 10 | 8.05 | |
R pars triangularis | 45 | 48 | 23 | 26 | 7.12 | |
R pars orbitalis | 47 | 45 | 32 | −6 | 7.46 | |
R middle frontal gyrus | 8 | 45 | 14 | 50 | 7.01 | |
L/R Pre-supplementary motor area (pre-SMA) | 8 | 234 | 6 | 32 | 46 | 8.78 |
−3 | 23 | 54 | 7.18 | |||
L anterior insular cortex (AIC) | 72 | −30 | 20 | −6 | 7.89 | |
R middle temporal gyrus | 21 | 24 | 57 | −43 | −2 | 5.62 |
21/22 | 7 | 54 | −25 | −6 | 5.41 | |
R dorsal premotor cortex | 6 | 9 | 21 | 14 | 50 | 5.40 |
GRAMMATICAL > UNGRAMMATICAL | ||||||
L/R ventromedial prefrontal cortex (vmPFC) | 11 | 80 | −12 | 41 | −10 | 7.22 |
6 | 41 | −14 | 6.27 | |||
TWO-LoE > ONE-LoE | ||||||
R middle frontal gyrus | 9 | 42 | 42 | 29 | 42 | 6.79 |
L inferior parietal lobule (IPL) | 39/40 | 31 | −54 | −49 | 42 | 6.31 |
R inferior parietal lobule (IPL) | 39 | 23 | 60 | −49 | 30 | 6.26 |
Pre-supplementary motor area (pre-SMA) | 8 | 21 | −6 | 32 | 34 | 5.85 |
R dorsal premotor cortex | 6 | 10 | 18 | 20 | 58 | 5.73 |
L middle frontal gyrus | 9 | 8 | −36 | 23 | 38 | 5.23 |
6/8 | 6 | −39 | 14 | 54 | 5.33 | |
L frontal pole | 10 | 5 | −33 | 53 | −2 | 5.32 |
ONE-LoE > TWO-LoE | ||||||
L supplementary motor area (SMA) | 6/4 | 151 | −36 | −19 | 58 | 6.90 |
L premotor cortex | 6 | 27 | −6 | 2 | 54 | 6.03 |
L superior temporal gyrus | 41 | 15 | −54 | −19 | 6 | 5.59 |
R superior temporal gyrus | 22 | 14 | 66 | −22 | 2 | 6.45 |
Notes: Reported clusters have a minimum cluster size of 4 voxels and were corrected for multiple comparisons voxel-wise using a threshold of p < 0.05. L – left hemisphere; R – right hemisphere; t-contrasts compared are shown in bold; Anatomical regions within each cluster are at least 4 mm apart and are ranked according to decreasing t-values.
For the main effect of LEVEL OF EMBEDDING (TWO-LoE > ONE-LoE), we observed bilaterally clusters in the inferior parietal lobule (IPL), middle frontal gyrus, dorsal premotor cortex, pre-supplementary motor area, and left frontal pole. For the reverse contrast (ONE-LoE > TWO-LoE), we identified clusters in the bilateral middle superior temporal gyri, and left premotor cortex.
No supra-threshold clusters were yielded for the interaction contrast (even at a more lenient cluster-wise FWE-corrected threshold of p < 0.05).
Results of the refined model which only contained category violations in ungrammatical sequences (see Materials and Methods) were furthermore analogous to the original model. The GRAMMATICALITY contrast (UNGRAMMATICAL > GRAMMATICAL) yielded four significant clusters with maxima in the right anterior insular cortex (coordinates in MNI space: [33, 26, 2]), right pars opercularis ([45, 17, 10]), right pars triangularis ([48, 23, 22]), and left anterior insular cortex (AIC; [−33, 20, −6]), whilst the LEVEL OF EMBEDDING contrast (TWO-LoE > ONE-LoE) yielded two significant clusters with maxima in the right inferior parietal lobule (IPL; [60, −49, 30]), and the right middle frontal gyrus ([42, 29, 42]). No significant clusters in the reverse and interaction contrasts were observed at the voxel-corrected statistical threshold (although a more-lenient cluster-corrected threshold nonetheless yielded significant clusters in the same regions as in the original model).
Psychophysiological Interaction (PPI) analysis
For the main effect of GRAMMATICALITY (UNGRAMMATICAL > GRAMMATICAL), we found significant psychophysiological interactions (see Fig. 3B and Table 4) between the right IPL (as seed) and the right pars opercularis and bilateral AIC. Only the psychophysiological interaction of the right IPL on the right AIC significantly correlated with participants’ overall sensitivity in discriminating between grammatical and ungrammatical sequences (see Fig. 3C, r = 0.55, p = 0.03, one-tailed test; corrected using Holm’s method across all significant clusters with the same seed region).
Table 4.
Anatomical region | BA | Cluster size (voxels) | MNI-coordinates | t-value | ||
---|---|---|---|---|---|---|
X | Y | Z | ||||
Seed: R inferior parietal lobule (IPL) | ||||||
R anterior insular cortex (AIC) | 9 | 30 | 29 | 2 | 4.34 | |
L anterior insular cortex (AIC) | 7 | −30 | 23 | −6 | 4.11 | |
R pars opercularis* | 44 | 9 | 51 | 20 | 22 | 3.87 |
Seed: R middle frontal gyrus | ||||||
R anterior insular cortex (AIC) | 9 | 30 | 26 | 2 | 4.63 | |
L anterior insular cortex (AIC) | 6 | −30 | 20 | −6 | 4.55 | |
R pars triangularis/ inferior frontal sulcus | 45/9 | 7 | 48 | 20 | 22 | 3.91 |
Notes: Reported clusters were corrected for multiple comparisons voxel-wise using a statistical threshold of p < 0.05 and an extent of 4 voxels. *This cluster had another local maxima 8.49 mm apart. L – left hemisphere; R – right hemisphere.
We also found significant interactions between the right middle frontal gyrus (as seed) and the right pars triangularis/inferior frontal sulcus and bilateral AIC with respect to GRAMMATICALITY Other combinations of contrasts and seed regions did not yield significant results.
Discussion
The present experiment aimed to uncover the functional neural basis underlying the human ability to process non-local dependencies – a key feature of hierarchical structures – in music. By independently manipulating the grammaticality and auditory tonal working memory demands of nested atonal musical sequences, we found that grammatical violations of nested musical dependencies led to increased BOLD responses in the right inferior frontal gyrus (IFG) and bilateral anterior insular cortices (AIC), whilst increased auditory working memory demands led to enhanced responses in the bilateral middle frontal gyri (MFG) and inferior parietal lobules (IPL). This result confirms our hypothesis that the inferior frontal gyrus – especially the right homologue of Broca’s area – is involved in processing non-local dependencies in music. Modulations in functional connectivity between these two distinct functional networks were also associated with discriminating between grammatical and ungrammatical nested sequences. In particular, the task-modulated connectivity between the right AIC and right IPL predicted behavioural performance. These suggest that resolving non-local dependencies in music requires the interplay between brain regions involved in processing hierarchical structures in music and brain regions involved in tonal working memory.
Our findings provide the first evidence that engagement of the right IFG in the neurocognition of music reflects processing of nested non-local dependencies based on internal knowledge of the grammatical rules of musical syntax. Importantly, our refined analysis suggests that the posterior right IFG is sensitive to violations of exclusively non-local dependencies in the absence of any local violation between immediate-adjacent elements. This clarifies prior work that had observed right-lateralised IFG activity (e.g. Koelsch et al., 2005, 2002) but employed paradigms that confounded hierarchical and local irregularity (e.g. a chord other than the tonic following a dominant seventh chord at the end of a chord sequence), such that observed IFG activity could have resulted from to local (serial) processing alone.
Participants’ post-experiment reports moreover suggest that the nested non-local dependencies were processed hierarchically. When prompted to explain what the underlying rule for the sequences were in an open-ended manner, all participants used terms such as ‘mirror’ (seven participants), ‘symmetry’ (four participants), ‘tree’ (one participant), or that the second half is the same as the first but in reversed/inverted order (five participants), to explain how the sequences were arranged. Given that each test sequence was unique and participants were never told what the generative rule was (see Materials and Methods), this demonstrates that participants were able to abstract surface features of the stimuli into a syntactic rule that describes their structure. This representation is moreover hierarchical because the A-motifs (e.g. A3A2A1) are grouped into a superordinate set that is mirror-transformed into another superordinate set containing the B-motifs (e.g. B1B2B3) in the second half. Importantly, participants exploited their understanding of the mirror-symmetric rule to accomplish the task. Participants actively predicted what the next motif could be in each sequence and compared their predictions with the incoming motif. A sequence was deemed grammatical if all predictions were met, and ungrammatical otherwise. This required participants to maintain multiple nested dependencies in parallel and in the correct order, as well as to monitor which superordinate class a motif belonged to (i.e. whether the upcoming motif would belong to the first or second half of a sequence).
Our finding of the right IFG in processing nested non-local dependencies in music thus supports the view that Broca’s area and its right homologue are involved in processing hierarchical structure, although with different hemispheric-weightings across various cognitive domains. While processing linguistic syntax is weighted towards the left hemisphere, we suggest that processing hierarchical structure in music is weighted towards the right: Previous studies have shown that the left IFG is sensitive to violations in nested artificial phoneme sequences64, and elicited increased activity for nested hierarchical sequences compared to serial non-hierarchical verbal sequences30,40,64,65. The pars opercularis of the left IFG in particular showed functional specificity to syntactic information in language29. Therefore, the observed activity in the right IFG – particularly the pars opercularis as highlighted in our refined analysis – likely reflects mechanisms which involve constructing hierarchical representations of the incoming acoustic information66.
The present data therefore raise the possibility that at least some aspects of hierarchical processing may rely on lateralised, domain-selective neuronal populations. Despite proposals for an interaction between syntactic processing in music and language32,34,67,68, our findings are more in line with the proposal that representing hierarchical dependencies could differ in music and language69, and thus calls for the existence of domain-specific resources in parsing the hierarchical dependencies. Musso and colleagues70, for example suggest that processing syntax in music and language is highly differentiated within the left IFG, but more generally engages in a dual-stream system that connects left frontal, parietal, and temporal regions. Our findings therefore motivate further studies on clarifying the extent to which music and language share common neural resources, and to what extent the observed effects in the right IFG pertain to syntactic processing specifically, or to general mechanisms such as attention71 or cognitive control72,73.
Nevertheless, our proposed clear-cut separation between the left IFG processing linguistic syntax and the right IFG processing musical syntax is only tentative: In light of the recent controversy on inflated false-positive rates in fMRI studies74, we adopted a conservative correction for multiple comparisons at the expense of statistical sensitivity – significantly increased BOLD responses to violations in musical syntax were indeed also observed in the left IFG at a more lenient cluster-wise corrected statistical threshold. Second, the participants of our study were trained musicians – a group shown to exhibit increased grey-matter volume in the right IFG75, higher fibre-tract volumes between the right IFG and the temporal lobe76, and differences in BOLD-response patterns77–79, compared to non-musicians. Alternatively, effects of lateralisation could have been driven by the idiosyncratic differences in experimental stimuli. For example, phonemes in speech are typically much shorter in duration than musical notes. It has been argued that auditory information is extracted at different timescales between the two hemispheres80, and that the left hemisphere is more specialised towards processing temporal features whilst the right towards spectral features of the auditory stream46,81. This mechanism could also explain how musical stimuli does not consistently engage the right or left IFG.
Furthermore, the role of hierarchy in resolving nested dependencies in artificial grammar studies has also been challenged. It has been argued that sequential processing is more parsimonious82, and that there is no intrinsic super-/subordinate relation across elements in a sequence83. In light of this, we are cautious in interpreting how processing nested non-local dependencies in our sequences extended towards hierarchical processing of music, and only suggest the role of the right IFG in the hierarchical processing of music based on participants’ self reports.
In addition to our experimental manipulation of grammaticality in a nested musical syntax, we also manipulated the demands of auditory tonal working memory by varying the number of embedding levels. When contrasting sequences with two levels of embedding against one, we observed decreased behavioural performance and increased BOLD responses in the inferior parietal lobule (IPL) and the middle frontal gyrus (MFG) of the dorsolateral prefrontal cortex (dlPFC). These effects likely reflect the established role of the IPL and dlPFC in tonal working memory, given the additional dependency induced by an additional pair of motifs to be held in memory. In accordance with this interpretation, activity in the MFG and IPL were shown to be involved in same/difference tasks in musical melodies46,84, and n-back tasks using chord sequences85. Similar responses have also been observed during the maintenance of pitch information43,45,46,86.
However, because motifs were matched with their transposed conjugates, participants had to maintain interval relations between the motifs and not their absolute pitch classes. Consequently, the contrast between sequences with two levels of embedding against one also reflects manipulation – on top of encoding and maintenance – of the incoming acoustic signals in tonal working memory. Previous studies have suggested that dlPFC activity may reflect modulatory top-down control signals to information represented elsewhere in the neocortex87–90, whilst the IPL was shown to be involved in integrating acoustic melodic information within a tonal context91, as well as comparing original and transposed melodies84,92. According to state-based models of working memory87,93,94, information held in working memory are distributed across the cortex and represented in increasing levels of abstractness: from relatively raw and unprocessed in the sensory cortices to highly abstract in the frontal cortex93. The observed increased BOLD response in the dlPFC may therefore reflect increased attention in maintaining the additional pair of motifs that is represented in a transposed and processed format in the IPL (please also see the Supplementary Information for the discussion on some brain areas for which we did not have clear a priori hypotheses).
Extending the functional main effects that we have just discussed, our psychophysiological interaction (PPI) analysis suggests that the behavioural sensitivity to violations of musical syntax requires the interaction of a system that processes nested hierarchical information (i.e. IFG and AIC), and an auditory working memory system (i.e. MFG and IPL). We thus propose that the ability to resolve nested hierarchies in music depends on how salient the constituent musical elements are represented in working memory. With this interpretation, BOLD response modulation in the AIC reflects participants’ awareness to violations that motivates the appropriate motor preparation and response. In addition to the involvement of the AIC in music cognition22,25,27,43,45,46,95,96, the right AIC is a key node of the salience network that detects behaviourally relevant stimuli97, supports the translation of affective signals into specific actions98,99 and is associated with perceptual decision-making, interoception, and emotional awareness98,100–103. A second mathematically equivalent interpretation of our PPI result is that processing violations in musical syntax instead drives functional connectivity between the IFG and AIC. We suggest that this is less plausible, because discriminating between grammatical and ungrammatical sequences required similar demands in maintaining the musical motifs in tonal working memory.
A limitation in comparing sequences with two levels of embedding against one is the inclusion of an extra pair of motifs that necessitates a longer sequence duration, which confounds distance and item effects. Although this limitation is present in other studies using similar artificial grammar-like nested sequences64,104, the effects of length were typically not observed, possibly because the differences in duration in those studies were around 1 s compared to 3 s in the current experiment. This difference could explain the supposed activity of the supplementary motor area when comparing sequences with one level of embedding against two – which most likely reflects motor preparation for the ensuing button response after the stimuli. Moreover, although we were able to infer from participants’ reports that they were sensitive to the order and pairing between A and B motifs, further studies could for example include a reordering violation (e.g. A3A2A1B3B1B2) to verify that participants were indeed sensitive to the order of motifs presented.
In conclusion, processing hierarchical structure in music may involve two functionally segregated but nonetheless interacting systems in the right hemisphere: the IFG and AIC in resolving nested non-local dependencies between musical elements, and the MFG and IPL in auditory tonal working memory. The resemblance of the right-lateralised activation profile in processing non-local dependencies in music with the left-hemispheric system in language moreover suggests that processing hierarchical structures in music and language involves qualitatively similar mechanisms subserved by domain-specific neural subpopulations.
Electronic supplementary material
Acknowledgements
This work was supported by the Max Plank Society (V.K.M.C., L.M., A.D.F., S.K.) and the Croucher Foundation (V.K.M.C.). We also thank all participants in the study, Chiao-Yi Wu, Mauricio Martins, Tomás Goucha, and Thomas C. Gunter for their helpful discussions, as well as the two anonymous reviewers for their contributions in improving this paper.
Author Contributions
V.K.M.C., S.K. and A.D.F. developed the paradigm. V.K.M.C. conducted the experiment. V.K.M.C. and L.M. analysed the data. V.K.M.C., L.M., A.D.F., and S.K. wrote and reviewed the manuscript.
Competing Interests
The authors declare no competing interests.
Footnotes
Electronic supplementary material
Supplementary information accompanies this paper at 10.1038/s41598-018-22144-9.
Publisher's note: Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
References
- 1.Schellenberg EG. Expectancy in melody: Tests of the implication-realization model. Cognition. 1996;58:75–125. doi: 10.1016/0010-0277(95)00665-6. [DOI] [PubMed] [Google Scholar]
- 2.Lerdahl, F. & Jackendoff, R. A Generative Theory of Tonal Music. (The MIT Press, 1983).
- 3.Rohrmeier M, Zuidema W, Wiggins GA, Scharff C. Principles of structure building in music, language and animal song. Philos. Trans. R. Soc. Lond. B. Biol. Sci. 2015;370:20140097. doi: 10.1098/rstb.2014.0097. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4.Rohrmeier M. Towards a generative syntax of tonal harmony. J. Math. Music. 2011;5:35–53. doi: 10.1080/17459737.2011.573676. [DOI] [Google Scholar]
- 5.Steedman MJ. A Generative Grammar for Jazz Chord Sequences. Music Percept. 1984;2:52–77. doi: 10.2307/40285282. [DOI] [Google Scholar]
- 6.Babbitt M. Twelve-tone invariants as compositional determinants. Music. Q. 1960;46:246–259. doi: 10.1093/mq/XLVI.2.246. [DOI] [Google Scholar]
- 7.Fitch WT, Martins MD. Hierarchical processing in music, language, and action: Lashley revisited. Ann. N. Y. Acad. Sci. 2014;1316:87–104. doi: 10.1111/nyas.12406. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Fitch WT. Toward a computational framework for cognitive biology: unifying approaches from cognitive neuroscience and comparative cognition. Phys. Life Rev. 2014;11:329–64. doi: 10.1016/j.plrev.2014.04.005. [DOI] [PubMed] [Google Scholar]
- 9.Badre D. Cognitive control, hierarchy, and the rostro-caudal organization of the frontal lobes. Trends Cogn. Sci. 2008;12:193–200. doi: 10.1016/j.tics.2008.02.004. [DOI] [PubMed] [Google Scholar]
- 10.Jeon H-A. Hierarchical processing in the prefrontal cortex in a variety of cognitive domains. Front. Syst. Neurosci. 2014;8:223. doi: 10.3389/fnsys.2014.00223. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11.Koelsch S, Rohrmeier M, Torrecuso R, Jentschke S. Processing of hierarchical syntactic structure in music. Proc. Natl. Acad. Sci. USA. 2013;110:15443–8. doi: 10.1073/pnas.1300272110. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.Martins MD, Gingras B, Puig-Waldmueller E, Fitch WT. Cognitive representation of ‘musical fractals’: Processing hierarchy and recursion in the auditory domain. Cognition. 2017;161:31–45. doi: 10.1016/j.cognition.2017.01.001. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.Tillmann B, Bigand E. Global context effect in normal and scrambled musical sequences. J. Exp. Psychol. Hum. Percept. Perform. 2001;27:1185–96. doi: 10.1037/0096-1523.27.5.1185. [DOI] [PubMed] [Google Scholar]
- 14.Bigand E, Pineau M. Global context effects on musical expectancy. Percept. Psychophys. 1997;59:1098–1107. doi: 10.3758/BF03205524. [DOI] [PubMed] [Google Scholar]
- 15.Kuhn G, Dienes Z. Implicit Learning of Nonlocal Musical Rules: Implicitly Learning More Than Chunks. J. Exp. Psychol. Learn. Mem. Cogn. 2005;31:1417–1432. doi: 10.1037/0278-7393.31.6.1417. [DOI] [PubMed] [Google Scholar]
- 16.Dienes Z, Longuet-Higgins C. Can musical transformations be implicitly learned? Cogn. Sci. 2004;28:531–558. doi: 10.1207/s15516709cog2804_2. [DOI] [Google Scholar]
- 17.Garza Villarreal EA, Brattico E, Leino S, Østergaard L, Vuust P. Distinct neural responses to chord violations: A multiple source analysis study. Brain Res. 2011;1389:103–114. doi: 10.1016/j.brainres.2011.02.089. [DOI] [PubMed] [Google Scholar]
- 18.Koelsch S, Gunter T, Friederici AD, Schröger E. Brain indices of music processing: ‘nonmusicians’ are musical. J. Cogn. Neurosci. 2000;12:520–541. doi: 10.1162/089892900562183. [DOI] [PubMed] [Google Scholar]
- 19.Maess B, Koelsch S, Gunter TC, Friederici AD. Musical syntax is processed in Broca’s area: an MEG study. Nat. Neurosci. 2001;4:540–545. doi: 10.1038/87502. [DOI] [PubMed] [Google Scholar]
- 20.Jentschke S, Friederici AD, Koelsch S. Neural correlates of music-syntactic processing in two-year old children. Dev. Cogn. Neurosci. 2014;9:200–208. doi: 10.1016/j.dcn.2014.04.005. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Loui P, Wu EH, Wessel DL, Knight RT. A Generalized Mechanism for Perception of Pitch Patterns. J. Neurosci. 2009;29:454–459. doi: 10.1523/JNEUROSCI.4503-08.2009. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22.Koelsch S, Fritz T, Schulze K, Alsop D, Schlaug G. Adults and children processing music: An fMRI study. Neuroimage. 2005;25:1068–1076. doi: 10.1016/j.neuroimage.2004.12.050. [DOI] [PubMed] [Google Scholar]
- 23.Tillmann B, Janata P, Bharucha JJ. Activation of the Inferior Frontal Cortex in Musical Priming. Ann. N. Y. Acad. Sci. 2003;999:209–211. doi: 10.1196/annals.1284.031. [DOI] [PubMed] [Google Scholar]
- 24.Tillmann B, et al. Cognitive priming in sung and instrumental music: Activation of inferior frontal cortex. Neuroimage. 2006;31:1771–1782. doi: 10.1016/j.neuroimage.2006.02.028. [DOI] [PubMed] [Google Scholar]
- 25.Vuust P, Roepstorff A, Wallentin M, Mouridsen K, Østergaard L. It don’t mean a thing…. Neuroimage. 2006;31:832–841. doi: 10.1016/j.neuroimage.2005.12.037. [DOI] [PubMed] [Google Scholar]
- 26.Minati L, et al. FMRI/ERP of musical syntax: comparison of melodies and unstructured note sequences. Neuroreport. 2008;19:1381–5. doi: 10.1097/WNR.0b013e32830c694b. [DOI] [PubMed] [Google Scholar]
- 27.Koelsch S, et al. Bach speaks: A cortical ‘language-network’ serves the processing of music. Neuroimage. 2002;17:956–966. doi: 10.1006/nimg.2002.1154. [DOI] [PubMed] [Google Scholar]
- 28.Meyer L, Obleser J, Anwander A, Friederici AD. Linking ordering in Broca’s area to storage in left temporo-parietal regions: The case of sentence processing. Neuroimage. 2012;62:1987–1998. doi: 10.1016/j.neuroimage.2012.05.052. [DOI] [PubMed] [Google Scholar]
- 29.Goucha T, Friederici AD. The language skeleton after dissecting meaning: A functional segregation within Broca’s Area. Neuroimage. 2015;114:294–302. doi: 10.1016/j.neuroimage.2015.04.011. [DOI] [PubMed] [Google Scholar]
- 30.Makuuchi M, Bahlmann J, Anwander A, Friederici AD. Segregating the core computational faculty of human language from working memory. Proc. Natl. Acad. Sci. USA. 2009;106:8362–7. doi: 10.1073/pnas.0810928106. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 31.Uddén J, Ingvar M, Hagoort P, Petersson KM. Broca’s region: A causal role in implicit processing of grammars with crossed non-adjacent dependencies. Cognition. 2017;164:188–198. doi: 10.1016/j.cognition.2017.03.010. [DOI] [PubMed] [Google Scholar]
- 32.Kunert R, Willems RM, Hagoort P. Language influences music harmony perception: effects of shared syntactic integration resources beyond attention. R. Soc. Open Sci. 2016;3:150685. doi: 10.1098/rsos.150685. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 33.Slevc LR, Rosenberg JC, Patel AD. Making psycholinguistics musical: Self-paced reading time evidence for shared processing of linguistic and musical syntax. Psychon. Bull. Rev. 2009;16:374–381. doi: 10.3758/16.2.374. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 34.Koelsch S, Gunter TC, Wittfoth M, Sammler D. Interaction between syntax processing in language and in music: an ERP Study. J. Cogn. Neurosci. 2005;17:1565–1577. doi: 10.1162/089892905774597290. [DOI] [PubMed] [Google Scholar]
- 35.Patel AD. Language, music, syntax and the brain. Nat. Neurosci. 2003;6:674–681. doi: 10.1038/nn1082. [DOI] [PubMed] [Google Scholar]
- 36.Tillmann B, Bigand E, Madurell F. Local versus global processing of harmonic cadences in the solution of musical puzzles. Psychol. Res. 1998;61:157–174. doi: 10.1007/s004260050022. [DOI] [Google Scholar]
- 37.Marvin EW, Brinkman A. The Effect of Modulation and Formal Manipulation on Perception of Tonic Closure by Expert Listeners. Music Percept. An Interdiscip. J. 1999;16:389–407. doi: 10.2307/40285801. [DOI] [Google Scholar]
- 38.Cook N. The Perception of Large-Scale Tonal Closure. Music Percept. 1987;5:197–205. doi: 10.2307/40285392. [DOI] [Google Scholar]
- 39.Konecni, V. J. In Advances in Pschology: Cognitive Processes in the Perception of Art (eds. Crozier, W. R. & Chapman, A. J.) 19, 71–93 (Elsevier Science Publishers B.V., 1984).
- 40.Bahlmann J, Schubotz RI, Friederici AD. Hierarchical artificial grammar processing engages Broca’s area. Neuroimage. 2008;42:525–534. doi: 10.1016/j.neuroimage.2008.04.249. [DOI] [PubMed] [Google Scholar]
- 41.Lai, J., Krahmer, E. & Sprenger, J. The learnability of Auditory Center-embedded Recursion. Proc. 37th Annu. Meet. Cogn. Sci. Soc. 1237–1242 (2015).
- 42.Lai J, Poletiek FH. The impact of adjacent-dependencies and staged-input on the learnability of center-embedded hierarchical structures. Cognition. 2011;118:265–273. doi: 10.1016/j.cognition.2010.11.011. [DOI] [PubMed] [Google Scholar]
- 43.Koelsch S, et al. Functional architecture of verbal and tonal working memory: An fMRI study. Hum. Brain Mapp. 2009;30:859–873. doi: 10.1002/hbm.20550. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 44.Kumar S, et al. A Brain System for Auditory Working Memory. J. Neurosci. 2016;36:4492–4505. doi: 10.1523/JNEUROSCI.4341-14.2016. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 45.Schulze K, Zysset S, Mueller K, Friederici AD, Koelsch S. Neuroarchitecture of verbal and tonal working memory in nonmusicians and musicians. Hum. Brain Mapp. 2011;32:771–783. doi: 10.1002/hbm.21060. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 46.Zatorre RJ, Evans AC, Meyer E. J. Neurosci. 1994. Neural mechanisms underlying melodic perception and memory for pitch; pp. 1908–1919. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 47.Wu, H. et al. Musical experience modulates categorical perception of lexical tones in native Chinese speakers. Front. Psychol. 6 (2015). [DOI] [PMC free article] [PubMed]
- 48.Steele CJ, Bailey JA, Zatorre RJ, Penhune VB. Early Musical Training and White-Matter Plasticity in the Corpus Callosum: Evidence for a Sensitive Period. J. Neurosci. 2013;33:1282–1290. doi: 10.1523/JNEUROSCI.3578-12.2013. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 49.Bernardi L. Cardiovascular, cerebrovascular, and respiratory changes induced by different types of music in musicians and non-musicians: the importance of silence. Heart. 2005;92:445–452. doi: 10.1136/hrt.2005.064600. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 50.Oldfield RC. The Assessment and Analysis of Handedness: The Edinburgh Inventory. Neuropsychologia. 1971;9:97–113. doi: 10.1016/0028-3932(71)90067-4. [DOI] [PubMed] [Google Scholar]
- 51.Müllensiefen, D., Gingras, B., Musil, J. & Stewart, L. The musicality of non-musicians: An index for assessing musical sophistication in the general population. PLoS One9 (2014). [DOI] [PMC free article] [PubMed]
- 52.Green, D. M. & Swets, J. A. Signal detection theory and psychophysics. (John Wiley & Sons, 1966).
- 53.Zhang J, Mueller ST. A note on ROC analysis and non-parametric estimate of sensitivity. Psychometrika. 2005;70:145–154. doi: 10.1007/s11336-003-1119-8. [DOI] [Google Scholar]
- 54.Lakens D. Calculating and reporting effect sizes to facilitate cumulative science: a practical primer for t-tests and ANOVAs. Front. Psychol. 2013;4:1–12. doi: 10.3389/fpsyg.2013.00863. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 55.de Vries MH, Monaghan P, Knecht S, Zwitserlood P. Syntactic structure and artificial grammar learning: The learnability of embedded hierarchical structures. Cognition. 2008;107:763–774. doi: 10.1016/j.cognition.2007.09.002. [DOI] [PubMed] [Google Scholar]
- 56.Perruchet P, Rey A. Does the mastery of center-embedded linguistic structures distinguish humans from nonhuman primates? Psychon. Bull. Rev. 2005;12:307–313. doi: 10.3758/BF03196377. [DOI] [PubMed] [Google Scholar]
- 57.Friston KJ, Holmes a, Poline JB, Price CJ, Frith CD. Detecting activations in PET and fMRI: levels of inference and power. Neuroimage. 1996;4:223–235. doi: 10.1006/nimg.1996.0074. [DOI] [PubMed] [Google Scholar]
- 58.Lacadie CM, Fulbright RK, Rajeevan N, Constable RT, Papademetris X. More accurate Talairach coordinates for neuroimaging using non-linear registration. Neuroimage. 2008;42:717–725. doi: 10.1016/j.neuroimage.2008.04.240. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 59.Eickhoff SB, et al. A new SPM toolbox for combining probabilistic cytoarchitectonic maps and functional imaging data. Neuroimage. 2005;25:1325–1335. doi: 10.1016/j.neuroimage.2004.12.034. [DOI] [PubMed] [Google Scholar]
- 60.O’Reilly JX, et al. Tools of the trade: psychophysiological interactions and functional connectivity. Soc. Cogn. Affect. Neurosci. 2012;7:604–9. doi: 10.1093/scan/nss055. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 61.McLaren DG, Ries ML, Xu G, Johnson SC. A generalized form of context-dependent psychophysiological interactions (gPPI): A comparison to standard approaches. Neuroimage. 2012;61:1277–1286. doi: 10.1016/j.neuroimage.2012.03.068. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 62.Friston KJ, et al. Psychophysiological and modulatory interactions in neuroimaging. Neuroimage. 1997;6:218–29. doi: 10.1006/nimg.1997.0291. [DOI] [PubMed] [Google Scholar]
- 63.Brett M, Anton J-L, Valabregue R, Poline J-B. Region of interest analysis using the MarsBar toolbox for SPM 99. Neuroimage. 2002;16:S497. [Google Scholar]
- 64.Friederici AD, Bahlmann J, Heim S, Schubotz RI, Anwander A. The brain differentiates human and non-human grammars: functional localization and structural connectivity. Proc. Natl. Acad. Sci. USA. 2006;103:2458–63. doi: 10.1073/pnas.0509389103. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 65.Friederici AD, Makuuchi M, Bahlmann J. The role of the posterior superior temporal cortex in sentence comprehension. Neuroreport. 2009;20:563–568. doi: 10.1097/WNR.0b013e3283297dee. [DOI] [PubMed] [Google Scholar]
- 66.Dehaene S, Meyniel F, Wacongne C, Wang L, Pallier C. The Neural Representation of Sequences: From Transition Probabilities to Algebraic Patterns and Linguistic Trees. Neuron. 2015;88:2–19. doi: 10.1016/j.neuron.2015.09.019. [DOI] [PubMed] [Google Scholar]
- 67.Fedorenko E, Patel AD, Casasanto D, Winawer J, Gibson E. Structural integration in language and music: evidence for a shared system. Mem. Cognit. 2009;37:1–9. doi: 10.3758/MC.37.1.1. [DOI] [PubMed] [Google Scholar]
- 68.Kunert, R., Willems, R. M., Casasanto, D. & Patel, A. D. Music and Language Syntax Interact in Broca’ s Area: An fMRI Study. 1–16, 10.1371/journal.pone.0141069 (2015). [DOI] [PMC free article] [PubMed]
- 69.Patel, A. D. & Morgan, E. Exploring Cognitive Relations Between Prediction in Language and Music. Cogn. Sci. 1–18, 10.1111/cogs.12411 (2016). [DOI] [PubMed]
- 70.Musso M, et al. A single dual-stream framework for syntactic computations in music and language. Neuroimage. 2015;117:267–283. doi: 10.1016/j.neuroimage.2015.05.020. [DOI] [PubMed] [Google Scholar]
- 71.Perruchet P, Poulin-Charronnat B. Challenging prior evidence for a shared syntactic processor for language and music. Psychon. Bull. Rev. 2013;20:310–7. doi: 10.3758/s13423-012-0344-5. [DOI] [PubMed] [Google Scholar]
- 72.Slevc LR, Okada BM. Processing structure in language and music: a case for shared reliance on cognitive control. Psychon. Bull. Rev. 2015;22:637–52. doi: 10.3758/s13423-014-0712-4. [DOI] [PubMed] [Google Scholar]
- 73.Fedorenko E, Behr MK, Kanwisher N. Functional specificity for high-level linguistic processing in the human brain. Proc. Natl. Acad. Sci. USA. 2011;108:16428–16433. doi: 10.1073/pnas.1112937108. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 74.Eklund A, Nichols TE, Knutsson H. Cluster failure: Why fMRI inferences for spatial extent have inflated false-positive rates. Proc. Natl. Acad. Sci. 2016;113:201602413. doi: 10.1073/pnas.1602413113. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 75.Sato, K., Kirino, E. & Tanaka, S. A Voxel-Based Morphometry Study of the Brain of University Students Majoring in Music and Nonmusic Disciplines. Behav. Neurol. 2015 (2015). [DOI] [PMC free article] [PubMed]
- 76.Halwani GF, Loui P, Rüber T, Schlaug G. Effects of Practice and Experience on the Arcuate Fasciculus: Comparing Singers, Instrumentalists, and Non-Musicians. Front. Psychol. 2011;2:1–9. doi: 10.3389/fpsyg.2011.00156. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 77.Vuust P, et al. To musicians, the message is in the meter. Neuroimage. 2005;24:560–564. doi: 10.1016/j.neuroimage.2004.08.039. [DOI] [PubMed] [Google Scholar]
- 78.Schlaug G, Jancke L, Huang Y, Steinmetz H. In vivo evidence of structural brain asymmetry in musicians. Science (80-.). 1995;267:699–701. doi: 10.1126/science.7839149. [DOI] [PubMed] [Google Scholar]
- 79.Jäncke L, Shah NJ, Peters M. Cortical activations in primary and secondary motor areas for complex bimanual movements in professional pianists. Cogn. Brain Res. 2000;10:177–183. doi: 10.1016/S0926-6410(00)00028-8. [DOI] [PubMed] [Google Scholar]
- 80.Poeppel D. The analysis of speech in different temporal integration windows: Cerebral lateralization as ‘asymmetric sampling in time’. Speech Commun. 2003;41:245–255. doi: 10.1016/S0167-6393(02)00107-3. [DOI] [Google Scholar]
- 81.Zatorre, R. J. Cerebral correlates of human auditory processing; perception of speech and musical sounds. Acoust. Signal Process. Cent. Audit. Syst. 453–468 (1997).
- 82.Frank SL, Bod R, Christiansen MH. How hierarchical is language use? Proc. R. Soc. B Biol. Sci. 2012;279:4522–4531. doi: 10.1098/rspb.2012.1741. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 83.Ojima S, Okanoya K. The Non-Hierarchical Nature of the Chomsky Hierarchy-Driven Artificial-Grammar Learning. BIOLINGUISTICS. 2014;8:163–180. [Google Scholar]
- 84.Jerde TA, Childs SK, Handy ST, Nagode JC, Pardo JV. Dissociable systems of working memory for rhythm and melody. Neuroimage. 2011;57:1572–1579. doi: 10.1016/j.neuroimage.2011.05.061. [DOI] [PubMed] [Google Scholar]
- 85.Pallesen KJ, et al. Cognitive Control in Auditory Working Memory Is Enhanced in Musicians. PLoS One. 2010;5:e11120. doi: 10.1371/journal.pone.0011120. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 86.Gaab N, Gaser C, Zaehle T, Jancke L, Schlaug G. Functional anatomy of pitch memory - An fMRI study with sparse temporal sampling. Neuroimage. 2003;19:1417–1426. doi: 10.1016/S1053-8119(03)00224-6. [DOI] [PubMed] [Google Scholar]
- 87.D’Esposito M, Postle BR. The Cognitive Neuroscience of Working Memory. Annu. Rev. Psychol. 2015;66:115–142. doi: 10.1146/annurev-psych-010814-015031. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 88.Curtis CE, D’Esposito M. Persistent activity in the prefrontal cortex during working memory. Trends Cogn. Sci. 2003;7:415–423. doi: 10.1016/S1364-6613(03)00197-9. [DOI] [PubMed] [Google Scholar]
- 89.Riggall AC, Postle BR. The Relationship between Working Memory Storage and Elevated Activity as Measured with Functional Magnetic Resonance Imaging. J. Neurosci. 2012;32:12990–8. doi: 10.1523/JNEUROSCI.1892-12.2012. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 90.Wager TD, Smith EE. Neuroimaging studies of working memory: a meta-analysis. Cogn. Affect. Behav. Neurosci. 2003;3:255–274. doi: 10.3758/CABN.3.4.255. [DOI] [PubMed] [Google Scholar]
- 91.Royal I, et al. Activation in the right inferior parietal lobule reflects the representation of musical structure beyond simple pitch discrimination. PLoS One. 2016;11:1–14. doi: 10.1371/journal.pone.0155291. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 92.Foster NEV, Zatorre R. J. A role for the intraparietal sulcus in transforming musical pitch information. Cereb. Cortex. 2010;20:1350–1359. doi: 10.1093/cercor/bhp199. [DOI] [PubMed] [Google Scholar]
- 93.Christophel TB, Klink PC, Spitzer B, Roelfsema PR, Haynes J-D. The Distributed Nature of Working Memory. Trends Cogn. Sci. 2017;21:111–124. doi: 10.1016/j.tics.2016.12.007. [DOI] [PubMed] [Google Scholar]
- 94.Ma WJ, et al. Changing concepts of working memory. Nat. Neurosci. 2014;17:347–56. doi: 10.1038/nn.3655. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 95.Platel H, et al. The structural components of music perception. A functional anatomical study. Brain. 1997;120(Pt 2):229–243. doi: 10.1093/brain/120.2.229. [DOI] [PubMed] [Google Scholar]
- 96.Lappe C, Steinsträter O, Pantev C. Rhythmic and melodic deviations in musical sequences recruit different cortical areas for mismatch detection. Front. Hum. Neurosci. 2013;7:260. doi: 10.3389/fnhum.2013.00260. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 97.Uddin LQ. Salience processing and insular cortical function and dysfunction. Nat. Rev. Neurosci. 2014;16:55–61. doi: 10.1038/nrn3857. [DOI] [PubMed] [Google Scholar]
- 98.Craig AD. How do you feel — now? The anterior insula and human awareness. Nat. Rev. Neurosci. 2009;10:59–70. doi: 10.1038/nrn2555. [DOI] [PubMed] [Google Scholar]
- 99.Wager TD, Barrett LF. From affect to control: Functional specialization of the insula in motivation and regulation. Emotion. 2004;129:2865. [Google Scholar]
- 100.Nieuwenhuys, R. The insular cortex. A review. Progress in Brain Research195, (Elsevier B.V., 2012). [DOI] [PubMed]
- 101.Chang LJ, Yarkoni T, Khaw MW, Sanfey AG. Decoding the role of the insula in human cognition: Functional parcellation and large-scale reverse inference. Cereb. Cortex. 2013;23:739–749. doi: 10.1093/cercor/bhs065. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 102.Kurth, F., Zilles, K., Fox, P. T., Laird, A. R. & Eickhoff, S. B. A link between the systems: functional differentiation and integration within the human insula revealed by meta-analysis. Brain Struct. Funct. 1–16, 10.1007/s00429-010-0255-z (2010). [DOI] [PMC free article] [PubMed]
- 103.Augustine JR. Circuitry and functional aspects of the insular lobe in primates including humans. Brain Res. Rev. 1996;22:229–244. doi: 10.1016/S0165-0173(96)00011-2. [DOI] [PubMed] [Google Scholar]
- 104.Bahlmann J, Schubotz RI, Mueller JL, Koester D, Friederici AD. Neural circuits of hierarchical visuo-spatial sequence processing. Brain Res. 2009;1298:161–70. doi: 10.1016/j.brainres.2009.08.017. [DOI] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.