Abstract
Humans are assumed to have a natural—universal—predisposition for making music and for musical interaction. Research in this domain is, however, typically conducted with musically trained individuals, and therefore confounded with expertise. Here, we present a rediscovered and updated invention—the E-music box—that we establish as an empirical method to investigate musical production and interaction in everyone. The E-music box transforms rotatory cyclical movements into pre-programmable digital musical output, with tempo varying according to rotation speed. The user’s movements are coded as continuous oscillatory data, which can be analysed using linear or nonlinear analytical tools. We conducted a proof-of-principle experiment to demonstrate that, using this method, pairs of non-musically trained individuals can interact according to conventional musical practices (leader/follower roles and lower-pitch dominance). The results suggest that the E-music box brings ‘active’ and ‘interactive’ musical capacities within everyone’s reach. We discuss the potential of this method for exploring the universal predisposition for music making and interaction in developmental and cross-cultural contexts, and for neurologic musical therapy and rehabilitation.
Keywords: music box, music production, joint action, universals, neurologic music therapy, non-musicians
1. Introduction
Current interest in the universality of music stems from the fact that virtually every human society makes—not just listens to—music [1]. The human predisposition to physically participate with music and to interact with others through music is observable early in life across a range of behaviours. These include infant–carer interactions as well as spontaneous movement to music, which starts in the foetus during the final stages of gestation [2] and develops through infancy [3,4] into more complex forms of action and interactions such as dance and collaborative music making [5].
However, research on musical universality has focused on music perception [6], with musical production and interaction receiving less attention. Furthermore, the relatively scant research on musical production is mostly conducted with musically trained individuals, and is therefore confounded with expertise with a given instrument, genre or musical system, as well as with the amount or effectiveness of musical training. These factors are difficult to control precisely. Therefore, the study of the human predisposition for music making could be revolutionized if the experimenters could bring the training approach, paradigm and exposure under their control. Individuals across a range of ages (infants, children, adults, aged people), cultures, cognitive skills and health conditions could thus be sampled without concerns over needing fine motor control of a conventional musical instrument.
Here, we present an empirical method for studying the human predisposition to make music—and to interact with others through music—in every person, and irrespective of musical training. We rediscovered a popular invention from the nineteenth century—the mechanical music box (figure 1a)—and updated it into an electromechanical device (figure 1b). The E-music box transforms cyclical rotatory movements into a musical melody whose tempo varies according to the velocity of the rotation. The tunes can be played with correct rhythm and constant tempo by rotating the handle with constant velocity. The E-music box’s digitalization allows for great flexibility in terms of input (movement) and output (music) parameters (which can be pre-programmed; figure 1d), and the generation of data with high spatio-temporal resolution. The user’s behaviour is coded both discretely (in terms of the produced effects) and continuously (in terms of the oscillatory movements), and can be analysed using a broad range of linear [7] and nonlinear [8] analytical tools.
The E-music box contributes to a growing arsenal of devices for exploring links between music and movement. Other devices addressing such links have been developed for a diverse range of purposes, such as selecting tracks from a playlist [9], experiencing musical agency [10] or measuring how people move in response to music [11]. Taking a different (but complementary) approach, the E-music box is designed specifically to allow everyone to use movement to make music, and to interact with others through music, as musicians—including ensemble musicians—do.
1.1. Proof of principle
As a proof of principle, we conducted an experiment in which pairs of non-musically trained individuals interacted with one another using E-music boxes. The two players controlled the tempo of two complementary parts from a popular song (‘Somewhere over the rainbow’; figure 2), and synchronized their musical outputs while hearing (but not seeing) each other’s performance.
We tested whether the non-musically trained pairs of participants would interact according to two common musical practices. First, we tested whether the players could intentionally co-perform according to leader–follower roles, resulting in the follower’s musical output lagging behind the leader’s output [12]. Second, we tested the widespread practice wherein lower-pitched voices dominate higher-pitched voices in conveying temporal information. This is commonly observed when bass-ranged instruments lay down the musical rhythm in an ensemble, which is likely to be due to superior time perception for lower musical pitch [13].
To this aim, the two players had control over two identical musical parts that differed only in terms of relative pitch: one player was assigned a higher pitch (high-pitch player) and the other a lower pitch (low-pitch player). Furthermore, the players were asked to synchronize their musical outputs while adopting leadership roles (leaders or followers) or spontaneously (no leader). When leadership roles were assigned, leaders were asked to dictate the overall tempo, and followers were asked to adapt to the leader’s tempo.
To the extent that our method provides a suitable testing ground for exploring universals in joint music making, we made two predictions. First, by computing measures of ‘instantaneous leadership’ (see Methods), we expected followers’ movements and musical outputs to lag behind leaders’ when leadership roles are assigned [12], whereas movements and musical outputs should be well synchronized in the absence of leadership roles. Second, by computing indexes of ‘coupling directionality’ (see Methods), we estimated which player (high pitch versus low pitch) influenced the other’s movements (and resultant musical tempo) to a stronger degree. It was expected that low-pitch players would influence the high-pitch players’ actions to a stronger degree than vice versa.
2. Methods
2.1. Participants
Sixteen pairs of participants (32 individuals, nine males) were recruited to participate in this study. Sample size was determined based on a small-to-medium effect of f=0.2 with α=0.05 and power=0.95 using G*Power [14], resulting in a required total sample size of nine pairs. Their age ranged between 18 and 41 years (mean 21.69, s.d.=6.16). None of these participants had ever received musical education or private musical training. All participants were right handed and none had hearing impairments. Ten participants were born in countries other than Australia, including Fiji (n=3), Pakistan (n=2), Bangladesh (n=1), Germany (n=1), Thailand (n=1), India (n=1) and Africa (unspecified nation, n=1). Participants’ average level of education was a high school certificate.
2.2. Musical material
For this study, we selected the first eight bars from the tune Somewhere over the rainbow, originally written for the 1939 movie The Wizard of Oz [15]. This tune was chosen for its broad popularity, which was assumed to facilitate the task for the non-musically trained participants. Familiarity with this song was rated 3.74 on average (s.d.=1.26) on a five-point Likert scale. Four participants who reported being unfamiliar with the song were presented with the original recording once (prior to commencing the experiment). The score of the tune (i.e. section used in this study) is presented in figure 2. The part higher in pitch (upper line in the score) and the part lower in pitch (lower line) were separated by two octaves (i.e. the fundamental frequency of the high-pitch tones was four times greater than the fundamental frequency of the low-pitch tones). The tone and register for the initial notes were C5 (higher-pitch part) and C3 (lower-pitch part). Musical instrument digital interface (MIDI) files were created for each note in the score (using the software Max/MSP, with piano as the musical instrument). Each note had equal duration (500 ms) and loudness (which was controlled by applying a constant velocity of 120). The MIDI files were then converted into wave files for compatibility with Presentation software (see below).
2.3. Hardware specifications of the E-music box
Two electromechanical E-music boxes were built by modifying DJ Heros, a low-cost music video game developed by FreeStyleGames to simulate turntables (figure 1b). The original circuitry from the turntables was removed and substituted with a microcontroller board (Arduino Leonardo, supplier Sparkfun Electronics) and Connectorized Transmissive Optoschmitt Sensors from Honeywell (HOA7720; supplier Jaycar Electronics) to create a high precision optical encoder. In addition, plastic handles (length=14 cm, diameter=2 cm) were attached on the turntables cover (6 cm from the centre). The handles were suited for gripping by the hand and were used to rotate the turntables (figure 1d). The two optoschmitt sensors, separated by 2 mm, were attached to the rotational platform of the turntable. The sensors were moved over the stationary encoder tooth train to produce two impulse trains phase shifted by ±90° (figure 1c).
The resolution of the rotary encoder described above is 0.8°. Output from the rotary encoder is processed by the microcontroller, which converts the pulse trains from the rotary encoder to an angular position with a resolution of 1.6°, which is incremented for clockwise rotation and decremented for counter-clockwise rotation. The output format for the microcontroller is a single byte in binary format (32-bit little-endian integer) counting from 0 to 255 (representing rotational increments of 1.6°). The Arduino board is fitted with a RS232 shield (Cutedigi Electronics, supplier Sparkfun Electronics) to provide a direct connection to the RS232 port on the computer. The serial port on the Arduino is configured with a baud rate of 9600, data rate of 960 bytes s−1, data bits—eight with no parity and one stop bit. The Arduino outputs data only when the plastic handle is turned, with maximal temporal resolution of 960 Hz. Although the hardware could support a higher resolution of 0.8°, the custom software (Presentation, Neurobehavioral Inc.) developed for the experiment is limited to receiving a single byte counting from 0 to 255 for efficiency. A detailed set of instructions on how to build an E-music box from a DJ Hero turntable, together with the Arduino code, are provided in the electronic supplementary material. The combined cost of building an E-music box is AUD $50, and all materials are available online (purchases made in 2014).
The E-music box could of course be built using different technology or hardware from ours. Yet, we believe that our implementation is very efficient to satisfy the conditions of (i) outputting data with high temporal and spatial resolution, (ii) minimizing costs, and (iii) relying on a movement that is simultaneously: (a) simple, natural and commonly widespread (in this case, similar to stirring a large spoon), (b) involving numerous joints and muscles (in this case, the upper limbs, resulting in a physically healthy activity that suits rehabilitation scenarios, see below) and (c) constrained in its trajectory (in this case, restricted by the trajectory of the handle) in order to obtain more controlled (i.e. less variable) data. For a discussion of why custom-built devices are generally preferable to commercially available solutions for empirical research, the reader is referred to Ravignani et al. [16].
2.4. Software specifications of the E-music box
The outputs from the two E-music boxes were fed into a high-performance Dell computer (Dell OptiPlex 960, with dual 3.0 GHz Xenon processors) via two separate serial ports. The computer was equipped with Presentation software, which was used for the delivery of the musical tones (as well as for controlling the experiment, see below). Presentation was programmed to collect the codes received from each E-music box, and to present series of WAV files corresponding to the two musical parts displayed in the score in figure 2. The codes received from one E-music box controlled the presentation timing of the tones belonging to only one musical part. Tone presentation timing depended on the succession of codes received as well as on the specific duration of each note specified in the musical score. In particular, it was necessary to rotate the E-music box handle by 45° in order to play a quaver, or eighth note (i.e. half a beat), by 90° for a crotchet, or quarter note (one beat) and by 180° for a minim, or half note (two beats; for a schematic illustration, see figure 2, bottom part, middle). This implied that, as long as the E-music box was rotated at constant velocity, the outputted music would preserve its metrical structure. This aspect, which was inspired by the mechanics of the original music box (figure 1a), was important in order to permit non-musically trained participants to produce metrically correct musical tunes. Presentation recorded the degree codes received from the E-music boxes, the codes’ timing and stimulus presentation.
2.5. Procedure
The two participants constituting a pair operated the E-music boxes from two separate soundproofed booths and therefore had no visual contact (figure 2). Each participant was seated at a desk equipped with a computer monitor, an E-music box and headphones (Sennheiser HD 280 pro).
Participants were instructed to rotate the E-music box with the right hand, and to synchronize their musical output with the musical output of their partner. They were told to synchronize, as accurately as possible, as if they were ‘singing a song together’. As a secondary goal, participants were asked to adopt leadership roles as requested (see below). Specifically, leaders were instructed to play at a pace that was at the same time comfortable for them, and achievable for the followers, who instead were instructed to adapt to the tempo of leaders. If ‘no leadership’ was requested, participants were asked to achieve synchronization spontaneously.
Each trial began with the visual presentation (on the computer monitor) of a fixation cross (500 ms), followed by a circular arrow (3000 ms). The arrow could either point in a clockwise or anticlockwise direction, which was meant to instruct the participants how to turn the E-music box handle during the trial. Next, a metronome playing four beats (750 ms intervals) was presented via headphones, while the monitor displayed an ‘ear’ symbol accompanied by text stating ‘listen’ (figure 2). When the metronome ended, a ‘go’ sign was presented on the monitor to indicate to start rotating the E-music box handle. From this point, by rotating the E-music box handle, each participant had control over the timing of the tones belonging to the musical part they were assigned to. Importantly, movement phase was coded relatively to the handle position at trial onset (i.e. the phase was set to zero when the ‘go’ sign was presented). The first outputted musical tone (consisting of a minim note; figure 2) was presented after a 45° rotation (in the instructed direction), regardless of the handle’s absolute position at trial onset (subsequent notes were produced following rotations of 45°, 90° or 180° depending on rhythmic duration, see above). The participants could always hear their musical output mixed with their partner’s output. A trial ended when both musical parts had been played entirely. At that point, the ‘go’ sign disappeared, and a new trial began (after a 500 ms intertrial interval).
All the pairs performed 64 trials, which were grouped into eight blocks. Each block comprised four trials during which the two participants turned the E-music box in the same direction and four in opposite directions. The assignment of directions was counterbalanced across participants. Leadership instructions were provided at the outset of each block, and remained unchanged for the whole block. These instructions were provided by presenting text on the monitor stating either: ‘you lead’ and ‘you follow’ (four blocks, with alternated instructions across participants) or ‘no leadership’ (four blocks, with the same instruction for both participants). The order of the blocks and the trials within each block was randomized. To familiarize participants with the procedure, each pair performed a few training trials before starting the experiment until they felt clear about the task. During this practice, it was particularly important to allow each participant to autonomously produce the part they were assigned to (higher or lower in pitch). This was crucial to permit the participants to discriminate self- versus other-produced musical outputs. For this reason, participants were randomly assigned to either the part lower or higher in pitch from the start of the practice, and this remained unchanged for the whole experimental session (which lasted around 40 min).
2.6. Data analysis
The collected data consisted of time series representing the onset time of the musical tones and the position of the E-music box handles (in degrees) within each trial. Representative data from one trial are provided in figure 3a,c, where the (rotatory) movement phases of each E-music box and the resultant musical tones are plotted as a function of time. These data were used to examine three properties of the interpersonal coordination: instantaneous leadership, coordination precision and coupling directionality. For each of these properties, two different measures were calculated.
2.6.1. Instantaneous leadership and coordination precision
The instantaneous leadership measures were intended to index which participant (high-pitch or low-pitch player) was lagging behind in the interpersonal coordination. We computed this index both discretely and continuously, i.e. at the level of the actual tones produced (lag of the discretely produced musical tones) as well as the movements employed to produce the musical tones (lag of the continuous rotatory movements).
Discrete data (onset time of the musical tones) were analysed consistently with sensorimotor synchronization research employing finger tapping or piano performance tasks [17,18]. Accordingly, we subtracted the onset time of each tone belonging to the low-pitch part from the complementary tone belonging to the high-pitch part. Next, we calculated the mean and the standard deviation of the computed signed asynchronies within each trial. Mean signed asynchronies indexed the instantaneous leadership. More specifically, positive mean asynchronies indicated that low-pitch tones were produced overall earlier than the high-pitch tones, while negative values indicated the opposite. The standard deviation of the asynchronies, on the other hand, indexed the precision of the coordination. The smaller the standard deviation, the more precise (i.e. less variable) is the coordination [18].
Continuous data, i.e. time series indicating the angular position of each E-music box handle in 1.6°-steps, were analysed consistently with previous research that examined interpersonal synchronization of oscillatory movements using continuous measures of relative phase [19]. Because the angular position of each E-music box handle was coded (in Presentation software) only when the handle was turned, sampling rate was not regular. Therefore, the positional time series were interpolated to form uniform time intervals of 10 ms (i.e. 100 Hz sampling rate), and all movements were converted into clockwise-like movements. The continuous phase of each participant corresponded to the angular position of his or her E-music box handle with respect to the position held at trial onset. Continuous relative phase was then computed by subtracting the unwrapped1 continuous phase of the high-pitch player from the unwrapped continuous phase of the low-pitch player [8] (shown in figure 3b). Accordingly, positive and negative relative phase values indicated that low pitch and high pitch led the coordination, respectively. Similar to the asynchrony data, we computed mean and standard deviation of the relative phase within each trial, to assess the instantaneous leadership and precision (i.e. variability) of the coordination.
2.6.2. Coupling directionality
The continuous phase time series were also used to estimate the directionality of the coupling within each trial, i.e. which participant (high-pitch or low-pitch player) in a pair was relatively more influential, and which one was more adaptive. For this purpose, we computed two indices.
First, we estimated the directionality of the coupling between participants using the evolution map approach proposed by Rosenblum & Pikovsky [20], also used in interpersonal coordination dynamics [21]. This method quantifies the directionality of the coupling between two self-sustained oscillators by estimating the ratio of the coupling terms from the phase (unwrapped) time series of the movements. This method provides a directionality index, signified d(1,2), which can vary from 1 to −1 (where 1 and −1 are the cases of unidirectional coupling, and 0 is the case of a perfectly symmetrical bidirectional coupling). Here, a positive value indicated higher influence of the low-pitch player over the high-pitch player, while a negative value indicated the opposite.
Second, we computed an additional measure indexing the directionality of the coupling, which was based on the assumption that leaders would be more regular in their movements, and followers more variable. Specifically, increasing movement variability was assumed to reflect the increasing adaptation (i.e. use of error correction) to achieve sensorimotor synchronization [22]. To compute this measure, we first estimated the velocity of rotation (separately for each participant) by computing the first-degree derivative of the unwrapped phases (figure 3d). This measure indicated the real-time relative movement velocity within a given trial (which increased and decreased according to the task’s demands; compare figure 3a and c with d). Next, we calculated the standard deviation of each participant’s velocity as an index of variability. Finally, we subtracted the low-pitch player variability from the high-pitch player variability to estimate whose velocity was more variable. This difference value, which we termed ‘relative timing variability’, was positive if the high-pitch player was more variable (i.e. more adaptive) than the low-pitch player, and negative if the low-pitch player was the more variable. Hence, positive relative timing variability indicated that the low-pitch player was influencing the high-pitch player more than vice versa.
2.7. Statistics
All dependent measures were analysed following the same procedure. First, trials during which one of the participants began rotating the handle in a direction different to the one instructed (during the first 2000 ms following the ‘go’ sign) or whose coordination index deviated more than 3 s.d. from the (participant-specific) mean were discarded (2.9% and 1.7% of the raw data, respectively). Next, condition-specific averages were computed separately for each pair of participants. A preliminary analysis determined that our dependent measures did not change as a function of whether the two participants were rotating the E-music box in the same direction (both clockwise or anticlockwise) or in different directions (one clockwise and the other anticlockwise). Therefore, these data were collapsed and entered into repeated measures ANOVAs with one within participants factor (leadership) having three levels: high-pitch leads, no leader, low-pitch leads.
3. Results
Our dependent measures of interpersonal coordination consisted of two indices of instantaneous leadership (mean signed asynchronies and relative phase), two indices of precision (standard deviation of the signed asynchronies and standard deviation of the relative phase) and, finally, two indices of coupling directionality (d(1,2) and relative timing variability; see ‘Data analysis’ section and figure 3).
3.1. Instantaneous leadership
Our measures of instantaneous leadership indexed which player (high-pitch or low-pitch) was lagging behind across conditions. These measures included mean signed asynchronies and mean relative phase, which can be seen in figure 4a and b, respectively. The ANOVAs on these data yielded a main effect of leadership for both the signed asynchronies (F2,30=13.84, p<0.00001) and the relative phase (F2,30=18.51, p<0.00001). These results indicated that instructed followers (both high-pitch and low-pitch players) produced their musical output slightly after the leaders (around 300 ms after, on average), and their E-music box handle lagged behind leaders’ handles (around 50° behind, on average). This result was further supported by significant one-sample t-tests comparing all high-pitch leads and low-pitch leads conditions versus zero (all ps<0.01).
During the no leader condition, the overall synchronization approached 0 ms asynchronies and zero relative phase degrees. This was confirmed by two separate one-sample t-tests comparing the asynchronies (t15=0.865, p=0.401) and the relative phase (t15=1.34, p=0.197) from the no leader conditions versus zero. This suggested that, when no leader was assigned, high- and low-pitch players were presumably involved in a state of instantaneous co-leadership.
3.2. Coordination precision
Our measures of coordination precision (see data in table 1) did not yield significant effects of leadership, neither for the standard deviation of the signed asynchronies (F2,30=1.097, p=0.347) nor the standard deviation of the relative phase (F2,30=2.602, p=0.091). The relative phase data, however, showed a trend towards significance. This was indicative of an overall increase in precision in the no leader condition compared with the conditions with a designed leader. Indeed, a planned comparison testing for the quadratic trend of the factor leadership (high-pitch and low-pitch versus no leader) reached significance (F2,30=7.325, p=0.016). Overall, these results indicated that, although precision tended to be higher in the no leader condition compared with the high-pitch and low-pitch conditions, the latter two were not different from one another. Thus, coordination precision did not change depending on whether low-pitch or high-pitch players led.
Table 1.
high-pitch leads | no leader | low-pitch leads | |
---|---|---|---|
asynchronies (s.d.) | 502.32±154.70 | 466.18±179.14 | 456.99±131.38 |
rel. phase (s.d.) | 77.30±26.62 | 68.16±26.90 | 72.07±21.64 |
3.3. Coupling directionality
The measures of coupling directionality—d(1,2) and relative timing variability—are displayed in figure 4c and d, respectively. The ANOVAs on these data yielded main effects of leadership for both d(1,2) (F2,30=13.197, p<0.0001) and relative timing variability (F2,30=15.704, p<0.0001). These results indicated that leadership instructions changed the directionality of the coupling between participants.
As predicted, both measures revealed an asymmetry in the coupling between low-pitch and high-pitch players. More specifically, these data indicated that—overall—low-pitch players influenced high-pitch players more than vice versa. This was demonstrated by t-tests contrasting each leadership condition versus zero. These tests revealed that, when the low-pitch players were leading, the d(1,2) (t15=5.147, p<0.001) and relative timing variability (t15=4.61, p<0.001) were significantly different from zero. This occurred also when no leadership roles were assigned (d(1,2): t15=2.85, p=0.012; relative timing variability: t15=2.30, p=0.036), but importantly, not when the high-pitch player was leading the coordination (d(1,2): t15=0.054, p=0.95; relative timing variability: t15=0.43, p=0.67).
Taken together, these results indicated that the coupling directionality was overall asymmetrical. Low-pitch players influenced high-pitch players to a stronger degree than vice versa, even when no leadership roles were designed. This suggests that controlling a relatively lower musical pitch led to stronger interpersonal influence over the partner.
4. Discussion
We presented a rediscovered and updated invention—the E-music box—that provides an empirical method to study the universal human predisposition to make music and to interact with others through music. Reinventing the wheel, the E-music box has the potential to revolutionize the scientific study of musical production by enabling rigorously controlled investigations in every person: across development, cultures, health conditions and irrespective of musical training.
In the following sections, we first discuss the results of the experiment we conducted (as a proof of principle) to demonstrate the potential of the E-music box to study musical interaction between individuals who never received musical training. Then, we discuss two research areas that would benefit from this empirical method: (i) the ontogenetic and phylogenetic origins of musical behaviour, and (ii) clinical interventions that use music for rehabilitation and therapy.
4.1. Proof of principle
A proof-of-principle experiment was conducted to assess the suitability of the E-music box to explore musical production and interaction, irrespective of musical training. To this aim, pairs of untrained participants performed a musical melody in ensemble (while hearing, but not seeing, each other) using E-music boxes.
By manipulating leadership roles (leader and follower) and relative pitch (one higher, one lower) of the interacting players, we observed two widespread musical practices. First, the players were able to take leader and follower roles when requested, which involved followers’ outputted music and movements lagging behind leaders’, as commonly observed in musical ensembles [12]. Furthermore, the participants were able to maintain near-perfect synchronization of the musical sounds when asked to synchronize spontaneously (no leader condition). This result is important because it demonstrates that non-musically trained individuals are potentially able—using this method—to achieve interpersonal synchronization during joint music making.
Second, besides addressing synchronization, our analyses explored to what extent the two players were able to mutually influence and adapt to each other’s actions. This was investigated using measures of coupling directionality. Consistent with the common musical practice where bass-ranged instruments lay down the musical rhythm [13], we observed an asymmetry in coupling directionality across high- and low-pitch players. Essentially, controlling a relatively lower musical pitch led to stronger interpersonal influence over the partner than vice versa, and this was observed even when none of the players was designed as leader.
These results highlight the potential of our empirical method to investigate musical interaction beyond musically trained individuals. Specifically, our proof of principle was successful in demonstrating that generally joint music making (implying synchronization and mutual adaptation), and particularly certain musical practices (such as leadership and lower-pitch dominance), can be effectively studied in the general population using our method.
4.2. Exploring the universal predisposition for music making and musical interaction
It is widely accepted that musical behaviour is a human universal. Yet, music’s origins and functions are mysterious. Theories span from music being a signal (a prelinguistic communication system [23]), to a tool to establish social bonds and group cohesion [24]. Similarly, research on human (ontogenetic) development suggests that music is largely a social endeavour [25]. When music and dance are social parts of devotional or ritual activity, they are participatory, with the possibility for all members to contribute in different ways: singing, playing an instrument, stamping one’s feet [26] or verbally interjecting [27]. Indeed, Malloch & Trevarthen [28] theorize that, from infancy, the core of human communication and companionship is ‘communicative musicality’ [28].
Despite the important social and interpersonal functions played by music across societies, the empirical study of musical universality has mostly focused on perception, such as perceived emotion [29] or emotion–movement correspondences [30]. Remarkably, musical production—without which there is no music to be perceived—and even more so musical interaction have been substantially neglected, perhaps owing to the confounding effects of expertise.
Here, we argue that the E-music box, by permitting every person to make familiar or unfamiliar, relatively intricate music and to interact with others through music, fills a gap in research. Indeed, given that the roots of music production are much older and that evolutionarily significant musical forms are communal and participatory, a musical device that enables joint music making without the need for extensive practice serves a significant research need. Moreover, a device that produces sounds that are not culturally constrained but able to be programmed to produce a range of timbres, scales and tunings is an asset for studying culturally diverse materials, populations, and in fieldwork settings.2 This enables rigorously controlled experiments that might eventually prove powerful in shedding light upon the mysterious ontogenetic [25] and phylogenetic [31,32] origins of music.
4.3. The E-music box for neurologic therapy and rehabilitation
Music-based movement activities have been successfully used as a complementary (low-cost) rehabilitation approach for treating acquired or congenital brain disorders, particularly those associated with motor dysfunctions [33]. These include gait-related activities in Parkinson’s disease [34] and stroke rehabilitation (including aphasia) [35].
However, music procedures that involve actual music making are scarcely used because they often require training and can only target patients with moderate impairments [36,37]. Here, we propose that the introduction of the E-music box in these therapeutic approaches could be particularly fruitful. Without undergoing any training procedure, a patient could play rewarding (i.e. patient-selected, [38]) musical tunes. Furthermore, input (movements) and output (music) parameters could be pre-programmed according to each patient’s needs. This could be a considerable advantage and eventually extend neurologic musical therapy to an easily accessible music making procedure that could be applied to a larger patient population than before.
The E-music box could introduce an interactive (yet very controlled) component in these procedures. The patients, besides making music, could possibly interact with another individual through music. Because the musical movements are coded as continuous oscillatory data, the entrainment between music players can be estimated using a range of powerful analytical tools—including linear and nonlinear techniques—that are suitable for assessing different aspects of performance [39,40]. This scenario could be particularly useful for traditional patient–therapist music therapy procedures and provide a methodologically sound method for testing the origins of its beneficial effects [41,42]. Furthermore, in the absence of a therapist, a computer-controlled virtual partner could operate the (interactive) E-music box (cf. [43,44]), and the degree of adaptivity of such a virtual partner could be progressively regulated according to the patient’s needs [45,46].
Finally, because pairs of E-music boxes can be used for studying interpersonal human interaction (as demonstrated here), this method is also likely to hold potential for understanding (and perhaps treating) social disorders where conventional communication may be impaired, such as autism [47]. This suggestion is reinforced by evidence indicating spared or enhanced musical abilities in autistic children [48], and musical interaction promoting empathic behaviour [49]. The call for novel interactive methods to study autism spectrum disorder [50], particularly musical interventions [51], is therefore answered by development of the E-music box.
5. Conclusion
We established an empirical method—the E-music box—to study the universal human predisposition for music making, and for interacting with others through music. The E-music box does not require training in order to be operated. Furthermore, its input (movement) and output (music) parameters can be flexibly pre-programmed. The E-music box thus permits virtually every person to produce diverse music and to interact with others through music in a controlled scenario. This opens interesting perspectives for research on the universal predisposition for music making, musical communication and musical interaction across ages (from infancy to elderly) and cultures. Moreover, this method might further develop the use of music for therapeutic and rehabilitation purposes, by bringing ‘active’ and ‘interactive’ musical practices within everyone’s reach.
Supplementary Material
Supplementary Material
Acknowledgements
The authors thank the MARCS tech team, Zoe Mitsopoulos and Kristin Geraci for their support throughout the development of the E-music box.
Footnotes
Phase unwrapping involves adding/subtracting 2π radians each time a discontinuity between 2π radians and zero radians (i.e. increasing after 360°, or decreasing after 0°) is detected (see [8]).
Besides providing more flexibility in overcoming constraints related to interindividual, intercultural or interage differences, it is also worth pointing out that the E-music box is a readily accessible method from the perspective of both the user (simple task) and experimenter (simple equipment to set up and no sophisticated analytical skills required).
Ethics
This study received ethical approval from the local (Western Sydney University) committee, and informed consent was obtained from each participant.
Data accessibility
Data are available from the Western Sydney University institutional repository: http://dx.doi.org/10.4225/35/561380fc38404.
Authors' contributions
G.N. conceived the method and study concept. All authors contributed to technical equipment development and study design. S.M. built the equipment and drafted the supplementary information. G.N. piloted the experiment and collected the data. G.N. analysed the data and interpreted the results with feedback from M.V. and P.E.K. G.N. drafted the paper, and M.V., P.E.K. and C.J.S. provided critical revisions. All authors approved the final version of the manuscript for submission.
Competing interests
P.E.K. is on the Royal Society Open Science Editorial board, but was not involved in the decision-making process for this manuscript.
Funding
The MARCS Institute for Brain, Behavior and Development and the Western Sydney University funded this research.
References
- 1.Cross I. 2001. Music, cognition, culture, and evolution. In The biological foundations of music: Annals of the New York Academy of Sciences (eds Zatorre RJ, Peretz I), pp. 28–42. New York, NY: The New York Academy of Sciences. [DOI] [PubMed] [Google Scholar]
- 2.Kisilevsky BS, Hains SMJ, Jacquet A-Y, Granier-Deferre C, Lecanuet JP. 2004. Maturation of fetal responses to music. Dev. Sci. 7, 550–559. (doi:10.1111/j.1467-7687.2004.00379.x) [DOI] [PubMed] [Google Scholar]
- 3.Zentner M, Eerola T. 2010. Rhythmic engagement with music in infancy. Proc. Natl Acad. Sci. USA 107, 5768–5773. (doi:10.1073/pnas.1000121107) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4.Phillips-Silver J, Trainor LJ. 2005. Feeling the beat: movement influences infant rhythm perception. Science 308, 1430 (doi:10.1126/science.1110922) [DOI] [PubMed] [Google Scholar]
- 5.D’Ausilio A, Novembre G, Fadiga L, Keller PE. 2015. What can music tell us about social interaction? Trends Cogn. Sci. 19, 1–4. (doi:10.1016/j.tics.2014.11.002) [DOI] [PubMed] [Google Scholar]
- 6.Stevens CJ. 2012. Music perception and cognition: a review of recent cross-cultural research. Top Cogn. Sci. 4, 653–667. (doi:10.1111/j.1756-8765.2012.01215.x) [DOI] [PubMed] [Google Scholar]
- 7.Repp BH, Su Y-H. 2013. Sensorimotor synchronization: a review of recent research (2006–2012). Psychon. Bull. Rev. 20, 403–452. (doi:10.3758/s13423-012-0371-2) [DOI] [PubMed] [Google Scholar]
- 8.Pikovsky A, Rosenblum M, Kurths J. 2003. Synchronization: a universal concept in nonlinear sciences. Cambridge, UK: Cambridge University Press. [Google Scholar]
- 9.Moens B, Noorden L, Van Leman M. 2010. D-jogger: syncing music with walking. In Proc. Sound and Music Computing Conf. 2010, Barcelona, Spain, pp. 451–456. See https://biblio.ugent.be/publication/1070528.
- 10.Fritz TH. et al 2013. Musical agency reduces perceived exertion during strenuous physical performance. Proc. Natl Acad. Sci. USA 110, 17 784–17 789. (doi:10.1073/pnas.1217252110) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11.Demey M, Leman M, Bossuyt F, Vanfleteren J. 2008. The musical synchrotron: using wireless motion sensors to study how social interaction affects synchronization with musical tempo. In 8th Int. Conf. on New Interfaces for Musical Expression NIME08, 2008. pp. 372–373.
- 12.Keller PE, Appel M. 2010. Individual differences, auditory imagery, and the coordination of body movements and sounds in musical ensembles. Music Percept. 28, 27–46. (doi:10.1525/mp.2010.28.1.27) [Google Scholar]
- 13.Hove MJ, Marie C, Bruce IC, Trainor LJ. 2014. Superior time perception for lower musical pitch explains why bass-ranged instruments lay down musical rhythms. Proc. Natl Acad. Sci. USA 111, 10 383–10 388. (doi:10.1073/pnas.1402039111) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Faul F, ErdFelder E, Lang A-G, Buchner A. 2007. A flexible statistical power analysis program for the social, behavioral, and biomedical sciences. Behav. Res. Methods 39, 175–191. (doi:10.3758/BF03193146) [DOI] [PubMed] [Google Scholar]
- 15.Harburg EY, Arlen H. 1939. Somewhere over the rainbow. From the film ‘The Wizard of Oz’.
- 16.Ravignani A, Matellán Olivera V, Gingras B, Hofer R, Rodríguez Hernández C, Sonnweber RS, Fitch W. 2013. Primate drum kit: a system for studying acoustic pattern production by non-human primates using acceleration and strain sensors. Sensors (Basel) 13, 9790–9820. (doi:10.3390/s130809790) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Keller PE, Novembre G, Hove MJ. 2014. Rhythm in joint action: psychological and neurophysiological mechanisms for real-time interpersonal coordination. Phil. Trans. R. Soc. B 369, 20130394 (doi:10.1098/rstb.2013.0394) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18.Repp BH. 2005. Sensorimotor synchronization: a review of the tapping literature. Psychon. Bull. Rev. 12, 969–992. (doi:10.3758/BF03206433) [DOI] [PubMed] [Google Scholar]
- 19.Schmidt RC, Carello C, Turvey MT. 1990. Phase transitions and critical fluctuations in the visual coordination of rhythmic movements between people. J. Exp. Psychol. Hum. Percept. Perform. 16, 227–247. (doi:10.1037/0096-1523.16.2.227) [DOI] [PubMed] [Google Scholar]
- 20.Rosenblum MG, Pikovsky AS. 2001. Detecting direction of coupling in interacting oscillators. Phys. Rev. E, Stat. Nonlin. Soft Matter Phys. 64, 045202 (doi:10.1103/PhysRevE.64.045202) [DOI] [PubMed] [Google Scholar]
- 21.Richardson MJ, Harrison SJ, Kallen RW, Walton A, Eiler BA, Saltzman E, Schmidt RC. 2015. Self-organized complementary joint action: behavioral dynamics of an interpersonal collision-avoidance task. J. Exp. Psychol. Hum. Percept. Perform. 41, 665–679. (doi:10.1037/xhp0000041) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22.Fairhurst MT, Janata P, Keller PE. 2014. Leading the follower: an fMRI investigation of dynamic cooperativity and leader–follower strategies in synchronization with an adaptive virtual partner. Neuroimage 84, 688–697. (doi:10.1016/j.neuroimage.2013.09.027) [DOI] [PubMed] [Google Scholar]
- 23.Mithen S. 2005. The singing Neanderthals: the origins of music, language, mind, and body. London, UK: Weidenfeld Nicolson. [Google Scholar]
- 24.Freeman WJI. 2000. A neurobiological role of music in social bonding. In The origins of music (eds Wallin N, Merkur B, Brown S), pp. 411–424. Cambridge, MA: MIT Press. [Google Scholar]
- 25.Trehub SE. 2003. The developmental origins of musicality. Nat. Neurosci. 6, 669–673. (doi:10.1038/nn1084) [DOI] [PubMed] [Google Scholar]
- 26.Barwick L. 2002. Tempo bands, metre and rhythmic mode in Marri Ngarr ‘Church Lirrga’ songs. Australas. Music Res. 7, 67–83. [Google Scholar]
- 27.Van der Meer W. 2014. Audience response and expressive pitch inflections in a live recording of legendary singer Kesar Bai Kerkar. In Expressiveness in music performance: empirical approaches across styles and cultures (eds D Fabian, R Timmers, E Schubert), pp. 170–184. Oxford, UK: Oxford University Press. [Google Scholar]
- 28.Malloch S, Trevarthen C. 2009. Communicative musicality: exploring the basis of human companionship. Oxford, UK: Oxford University Press. [Google Scholar]
- 29.Fritz T, Jentschke S, Gosselin N, Sammler D, Peretz I, Turner R, Friederici AD, Koelsch S. 2009. Universal recognition of three basic emotions in music. Curr. Biol. 19, 573–576. (doi:10.1016/j.cub.2009.02.058) [DOI] [PubMed] [Google Scholar]
- 30.Sievers B, Polansky L, Casey M, Wheatley T. 2013. Music and movement share a dynamic structure that supports universal expressions of emotion. Proc. Natl Acad. Sci. USA 110, 70–75. (doi:10.1073/pnas.1209023110) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 31.Hauser MD, McDermott J. 2003. The evolution of the music faculty: a comparative perspective. Nat. Neurosci. 6, 663–668. (doi:10.1038/nn1080) [DOI] [PubMed] [Google Scholar]
- 32.Fitch WT. 2006. On the biology and evolution of music. Music Percept. 24, 85–88. (doi:10.1525/mp.2006.24.1.85) [Google Scholar]
- 33.Altenmüller EO, Schlaug G. 2015. Apollo’s gift: new aspects of neurologic music therapy. Prog. Brain Res. 217, 237–252. (doi:10.1016/bs.pbr.2014.11.029) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 34.De Dreu MJ, van der Wilk ASD, Poppe E, Kwakkel G, van Wegen EEH. 2012. Rehabilitation, exercise therapy and music in patients with Parkinson’s disease: a meta-analysis of the effects of music-based movement therapy on walking ability, balance and quality of life. Parkinsonism Relat. Disord. 18, S114–S119. (doi:10.1016/S1353-8020(11)70036-0) [DOI] [PubMed] [Google Scholar]
- 35.Thaut MH, McIntosh GC. 2014. Neurologic music therapy in stroke rehabilitation. Curr. Phys. Med. Rehabil. Rep. 2, 106–113. (doi:10.1007/s40141-014-0049-y) [Google Scholar]
- 36.Amengual JL. et al. 2013. Sensorimotor plasticity after music-supported therapy in chronic stroke patients revealed by transcranial magnetic stimulation. PLoS ONE 8, 1–10. (doi:10.1371/journal.pone.0061883) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37.Schneider S, Schönle PW, Altenmüller E, Münte TF. 2007. Using musical instruments to improve motor skill recovery following a stroke. J. Neurol. 254, 1339–1346. (doi:10.1007/s00415-006-0523-2) [DOI] [PubMed] [Google Scholar]
- 38.Salimpoor VN, Benovoy M, Larcher K, Dagher A, Zatorre RJ. 2011. Anatomically distinct dopamine release during anticipation and experience of peak emotion to music. Nat. Neurosci. 14, 257–262. (doi:10.1038/nn.2726) [DOI] [PubMed] [Google Scholar]
- 39.Wing AM, Endo S, Bradbury A, Vorberg D. 2014. Optimal feedback correction in string quartet synchronization. J. R. Soc. Interface 11, 20131125 (doi:10.1098/rsif.2013.1125) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 40.Walton AE, Richardson MJ, Langland-Hassan P, Chemero A. 2015. Improvisation and the self-organization of multiple musical bodies. Front. Psychol. 6, 313 (doi:10.3389/fpsyg.2015.00313) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 41.Koelsch S. 2009. A neuroscientific perspective on music therapy. Ann. N.Y. Acad. Sci. 1169, 374–384. (doi:10.1111/j.1749-6632.2009.04592.x) [DOI] [PubMed] [Google Scholar]
- 42.Thaut M. 2014. Neurobiological foundations of neurologic music therapy: rhythmic entrainment and the motor system. Front. Psychol. 5, 1185. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 43.Repp BH, Keller PE. 2008. Sensorimotor synchronization with adaptively timed sequences. Hum. Mov. Sci. 27, 423–456. (doi:10.1016/j.humov.2008.02.016) [DOI] [PubMed] [Google Scholar]
- 44.Van der Steen MCM, Keller PE. 2013. The adaptation and anticipation model (ADAM) of sensorimotor synchronization. Front. Hum. Neurosci. 7, 253 (doi:10.3389/fnhum.2013.00253) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 45.Hove MJ, Keller PE. 2015. Impaired movement timing in neurological disorders: rehabilitation and treatment strategies. Ann. N.Y. Acad. Sci. 1337, 111–117. (doi:10.1111/nyas.12615) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 46.Hove MJ, Suzuki K, Uchitomi H, Orimo S, Miyake Y. 2012. Interactive rhythmic auditory stimulation reinstates natural 1/F timing in gait of Parkinson’s patients. PLoS ONE 7, e32600. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 47.Simpson K, Keen D. 2011. Music interventions for children with autism: narrative review of the literature. J. Autism Dev. Disord. 41, 1507–1514. (doi:10.1007/s10803-010-1172-y) [DOI] [PubMed] [Google Scholar]
- 48.Heaton P. 2009. Assessing musical skills in autistic children who are not savants. Phil. Trans. R. Soc. 364, 1443–1447. (doi:10.1098/rstb.2008.0327) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 49.Rabinowitch TC, Cross I, Burnard P. 2013. Long- term musical group interaction has a positive influence on empathy in children. Psychol. Music 41, 484–498. (doi:10.1177/0305735612440609)) [Google Scholar]
- 50.Rolison MJ, Naples AJ, Mcpartland JC. 2015. Interactive social neuroscience to study autism spectrum disorder. Yale J. Biol. Med. 88, 17–24. [PMC free article] [PubMed] [Google Scholar]
- 51.Greenberg DM, Rentfrow PJ, Baron-Cohen S. 2015. Can music increase empathy? Interpreting musical experience through the empathizing—systemizing (E-S) theory: implications for autism. Empir Musicol Rev. 10, 79–94. [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
Data are available from the Western Sydney University institutional repository: http://dx.doi.org/10.4225/35/561380fc38404.