Skip to main content
PLOS One logoLink to PLOS One
. 2020 Feb 6;15(2):e0228457. doi: 10.1371/journal.pone.0228457

Do the shuffle: Exploring reasons for music listening through shuffled play

Katie Rose M Sanfilippo 1, Neta Spiro 2,3,*, Miguel Molina-Solana 3,4, Alexandra Lamont 5
Editor: Sarah EP Munce6
PMCID: PMC7004375  PMID: 32027672

Abstract

Adults listen to music for an average of 18 hours a week (with some people reaching more than double that). With rapidly changing technology, music collections have become overwhelmingly digital ushering in changes in listening habits, especially when it comes to listening on personal devices. By using interactive visualizations, descriptive analysis and thematic analysis, this project aims to explore why people download and listen to music and which aspects of the music listening experience are prioritized when people talk about tracks on their device. Using a newly developed data collection method, Shuffled Play, 397 participants answered open-ended and closed research questions through a short online questionnaire after shuffling their music library and playing two pieces as prompts for reflections. The findings of this study highlight that when talking about tracks on their personal devices, people prioritise characterizing them using sound and musical features and associating them with the informational context around them (artist, album, and genre) over their emotional responses to them. The results also highlight that people listen to and download music because they like it–a straightforward but important observation that is sometimes glossed over in previous research. These findings have implications for future work in understanding music, its uses and its functions in peoples’ everyday lives.

Introduction

Music listening is a ubiquitous and constant phenomenon in industrialised society. Adults listen to music for an average of 18 hours a week (with some participants reaching more than double that) [1]. Music is heard between 44% and 68% of people’s waking hours, accompanying a range of activities such as travel, eating, exercise, work and study [24]. While music is found in many public settings including transport, shops, restaurants and the workplace, self-chosen individual music listening represents an important aspect of musical engagement. North et al. [3] reported 37% of music listening episodes were undertaken alone, and of these, 82% were episodes of self-chosen music. Similarly, Greasley and Lamont [1] found 55% of music listening experiences were when participants were alone. Emotional reactions to music have been shown to be stronger when music is chosen by its listeners [4], and the increase in availability and portability of personal listening devices suggests that people are now able to choose music and curate music libraries in an increasing range of situations and for a range of activities.

Building on research from experimental aesthetics investigating the fit between music and context, recent research has begun to explore the role of situation. For instance, North and Hargreaves [5] found that listeners preferred high arousal versions of a piece while exercising and low arousal versions of the same piece while relaxing. More recently, researchers have investigated both the contexts and the activities that are associated with music, finding differences in both [1,3,6]. For instance, music in the gym is found to be more motivating than music in a restaurant [6]. Greb, Schlotz and Steffens [7] directly compared the effects of individual and situational influences on different functions of music listening, asking participants to imagine and then describe self-chosen listening situations. They found situational influences to be more important (see also [8]), concluding that active engagement with music can only be understood when accounting for context.

Similarly, in terms of functions of music listening, research has identified a range of motivations for listening [9] and functions that music serves [10]. A key motivation for (and outcome of) music listening is its emotional response [9,1113]. Other motivations include social relationships and connections, seeking information about music, and music’s potential for movement (tapping along or dancing) [9,14]. Music serves a range of functions including intellectual stimulation, coping with problems, motor synchronisation and wellbeing, overcoming loneliness, distraction, and defining and marking personal and social identity [7,13,14]. The pleasure of music listening has been found to result from the motives of relaxation, power, joy, and kinship [15].

Functions and uses have also been linked to aspects of the music and music preference. In an experimental study Sallavanti, Szilagyi and Crawley [16] found listeners used more complex music for cognitive purposes and less complex music for emotional purposes. Schäfer and Sedlmeier [17] found people who had strong preferences for their favourite music, whatever it might be, also emphasised the communicative functions of music listening such as expressing one’s own identity (cf. [18]), and tended to show more functions of their listening behaviour (cf. [1]). While mood, arousal and emotional benefits were common across all styles, they also found that fans of different styles emphasised different functions of their own preferred music. For instance, electronic music fans appreciated the sense of energy and ability of the music to put them in a good or ecstatic mood, along with understanding one’s own thoughts and feelings, while pop music fans liked the music’s ability to express their values and the sense of identification with artists and other people.

Much of the research reviewed so far has either presented participants with experimenter-selected music or asked people to report preferred styles, recreate memories of their favourite music, or talk about the music they like. One main drawback of these approaches is that participants either have no choice of musical stimulus or they are relying on memory which, although sometimes vivid [19,20], is not always accurate [21]. A more robust way of exploring preference is to focus on people’s own music collections. In studies that have done this, participants have been invited to select, discuss, and even play their favourite music during an interview [22]. However, even in an interview study that does this (e.g. Greasley, Lamont and Sloboda [22]) participants’ favourite music was emphasised, thus excluding their perceptions of music in their collection that may not be their favourite or even music they like.

Rapidly changing technology means the physical collection as studied by Greasley, Lamont, and Sloboda [22] in the mid-2010s has become overwhelmingly digital. Krause, North, and Hewitt [23] found that young adults now carry most of their music with them on personal listening devices; and thus data from playlists is beginning to be deliberately used in research. For instance, Krause and North [24] asked participants to create playlists for specific situations, confirming earlier findings that music choices can be heavily influenced by situation. This reflects a more curated approach to research than experience sampling, which aims to capture random slices of listening experiences. Between these two poles, random or ‘shuffle’ modes of listening provide an interesting midpoint between control and chance. Shuffle provides a more flexible approach to musical engagement [24], and is a very popular listening strategy, as part of the new flexibility that technology affords [25]. For instance, 35% of the mobile participants in Heye and Lamont's [26] study were listening via shuffle mode (the largest response category) while travelling.

Aims and research questions

The current study bridges the gap between reflective interviews about curated self-chosen music and random experience sampling of music listening in everyday settings. The aim of this project is to investigate which aspects of the music listening experience are prioritized by people listening to music on their personal devices. We do so by developing a new data collection method, Shuffled Play. This method does not specifically focus on people’s favourite music or music played in the moment. Instead it starts from a quasi-random selection of music that people have chosen as part of their personal music libraries which they access using the shuffle function. With this method, along with collecting mixed response type data, we build on and go beyond the more typical foci in the literature on music and emotion, music in everyday life and music and technology.

In this study we specifically explore two main research questions through a short online questionnaire asking participants to shuffle their music library and play two tracks as a prompt for reflection:

  1. Why do people download and listen to music?

  2. Which aspects of the music listening experience are prioritized when people talk about a track on their device?

Materials and methods

Design

For this descriptive study, we collected responses from a mix of question response types, including closed and open questions. We administered a short online questionnaire asking participants to shuffle their music library on their listening device and answer open and closed questions about the two music tracks that were selected first. Ethical approval was obtained from the Nordoff Robbins Ethics Committee and Keele University. All written consent was obtained online through the online questionnaire.

Participants

398 people began the survey and 397 completed information regarding music on their device. Responses from these 397 participants were used in the analysis of participants’ open-ended responses to their music. Demographic information was completed by 322 participants. About two thirds of the 322 participants identified as female (n = 202, 63%) and all were between the ages of 18 and 66 years old (M = 30.15, SD = 11.36). Sixty-nine percent of the participants were aged between 18 and 35 (n = 223, 69%). The majority were from, or currently resided in, the UK (n = 226, 83% and n = 257, 79.5% respectively) (Table 1).

Table 1. Demographic characteristics of the 322 participants.

n (%) out of 322
How do you describe yourself?
Female 202 (62.73)
Male 117 (36.33)
Would rather not say 3 (0.93)
How old are you?
18–24 145 (45.03)
25–34 78 (24.22)
35–44 56 (17.39)
45–54 26 (8.07)
55–64 16 (5.00)
65–74 1 (0.31)
Which country are you originally from?
UK 226 (70.18)
Bulgaria 3 (0.93)
Denmark 1 (0.31)
Finland 1 (0.31)
France 2 (0.62)
Germany 6 (1.86)
Greece 2 (0.62)
Ireland 9 (2.80)
Netherlands 6 (1.86)
Poland 2 (0.62)
Portugal 1 (0.31)
Slovakia 1 (0.31)
Slovenia 1 (0.31)
Spain 2 (0.62)
Sweden 2 (0.62)
Switzerland 1 (0.31)
Turkey 2 (0.62)
Total European countries 268 (83.23)
Canada 5 (1.55)
USA 30 (9.32)
Total North America 35 (10.87)
Brazil 2 (0.62)
Brunei 2 (0.62)
China 1 (0.31)
Ghana 1 (0.31)
Kenya 1 (0.31)
New Zealand 1 (0.31)
Nigeria 2 (0.62)
Pakistan 1 (0.31)
Peru 1 (0.31)
Philippines 2 (0.62)
Singapore 1 (0.31)
South Africa 2 (0.62)
Zambia 1 (0.31)
Zimbabwe 1 (0.31)
Total other countries 19 (5.90)
In which country do you currently reside?
UK 256 (79.50)
Belgium 2 (0.62)
Czech Republic 1 (0.31)
Denmark 1 (0.31)
Finland 1 (0.31)
France 1 (0.31)
Germany 3 (0.93)
Greece 2 (0.62)
Ireland 6 (1.86)
Latvia 1 (0.31)
Netherlands 1 (0.31)
Poland 1 (0.31)
Slovenia 1 (0.31)
Spain 1 (0.31)
Sweden 2 (0.62)
Total European countries 280 (86.96)
Canada 4 (1.24)
Mexico 1 (0.31)
United States of America 30 (9.32)
Total North America 35 (10.87)
Algeria 1 (0.31)
Australia 1 (0.31)
Brazil 1 (0.31)
New Zealand 2 (0.62)
Singapore 1 (0.31)
South Africa 1 (0.31)
Total other countries 7 (2.17)

Note. Only 70 out of the 322 answered the question “Why do you usually listen to this track?” (described below) due to a glitch in the survey program. No significant differences in demographic information were found between those that answered this question and those that did not.

As Table 2 details, the vast majority of the participants did not use music in their professional life (n = 227, 70%). The average score on the Active Engagement Subscale of the Goldsmiths Music Sophistication Index was 40.50 (SD = 8.19) corresponding to the 43rd percentile of the data norm reported in Müllensiefen et al. [27]. Our sample was thus representative of the average level of musical engagement. The vast majority of participants usually listened to music on their personal listening devices (n = 281, 87%). The two most popular ways of choosing music to listen to were by using self-selected playlists (n = 119, 30%) and choosing particular tracks (n = 146, 45%).

Table 2. Musical experience of the 322 participants.

n (%) for 322
I use music in my professional life
Yes 95 (29.50)
No 277 (70.50)
Music Sophistication Index: Active Engagement Subscale
10–19 2 (0.62)
20–29 26 (8.07)
30–39 106 (32.92)
40–49 141 (43.79)
50–59 47 (14.60)
How do you usually listen to your music?
On my personal listening devices (through iTunes, Spotify, Soundcloud, etc.) 281 (87.27)
Other 41 (12.73)
How do you usually choose the music you listen to?
Choose which tracks I want to listen to when I want to listen to them 146 (45.34)
Use self-selected playlists 119 (36.96)
Using playlists created by others 18 (5.59)
Radio 15 (4.66)
Other 24 (7.45)
I usually . . . (select all that apply) n (%) for 414
Buy my music online 116 (28.02)
Buy my music in a shop 11 (2.65)
Listen via a subscription to a site (e.g. Spotify) 194 (46.86)
Get my music for free 81 (19.56)
Other 12 (2.90)
Total 414

Note. Only 70 out of the 322 answered the question “Why do you usually listen to this track?” (described below) due to a glitch in the survey program. No significant differences in musical experience were found between those that answered this question and those that did not. “I usually…” was a select all that apply. Therefore the total % is out of 414 total options that were chosen.

A new data collection method: Shuffled play

The online questionnaire consisted of three sections. In the first section, participants were asked to use their regular music listening device (e.g. phone, laptop, etc.), turn on the shuffle function within a listening app they usually use (e.g. Spotify, iTunes, etc.) and press play. Whatever track came on first, they were asked to provide the track title, artist, device and app they were using. They were then asked open-ended questions about this track: “What's the first thing that comes into your mind about this track?” and “Why did you choose to save this track to your device?”. They were asked whether they enjoyed the track on a five point scale from Really enjoy = 5 to Really do not enjoy = 1 followed by an open question: “Why do you enjoy or not enjoy listening to the track?”. They were asked about their relationship with the track: “Do you have a relationship with this track? Whether it's 'Yes' or 'No' please explain”. The participant then pressed shuffle again and repeated this section about a new track. Once they had responded to all open-ended questions for both tracks they went onto the second section of the questionnaire.

The second section included three multiple-choice questions about the tracks they had just written about. Participants were presented with all the closed questions to be answered in relation to the first track they discussed in the previous section. They were reminded of the track title and artist in the questionnaire. They then answered all the questions about the second track they had discussed in the previous section, again being reminded of the track’s title and artist. The first, “Why do you usually listen to this track?”, was based on the 16 statements used by Greasley and Lamont [1]. Unlike in the original paper where they were asked to choose all that apply, each participant was asked to rank the reasons from 1–5 (1 = most relevant, 5 = least relevant). Due to a technical glitch, only 70 participants (20% of the participants) filled out this question. As we were not aware of an existing set of questions that had been used to specifically understand why people download music and save it rather than listen to it, we created our own question, “Why did you download this track?”, and provided 15 options covering three areas: (1) they liked it or something about it, (2) they heard it somewhere or (3) they were recommended it by someone else or it had meaning for them. Participants could tick all that applied of these options. Finally, we asked when they had last heard the tracks.

The third section included questions about their music listening habits and demographic characteristics. They only answered these questions once. To measure musical engagement we used the Goldsmiths Musical Sophistication Index, Musical Engagement Subscale [28]. This nine-item scale (α = .872) measures the amount of musical engagement or the amount of time and money people spend on music or music-related activity.

The questionnaire closed with two items that checked how participants had completed the survey. We asked whether they had skipped any tracks before they settled on the track to answer about and whether they had listened to the tracks while completing the questionnaire. Finally, participants could provide comments or feedback about the survey.

The survey, that can be found in full in the supporting information (S1 File), was created and administered online using Qualtrics [29].

Procedure

People with and without musical training were invited to participate via social media sites and over the internet. A snowball sampling [30,31] recruitment procedure was used with direct email invitations sent to target participants from the authors' professional and social networks. This allowed the researchers to specifically target different populations, i.e. international and older participants, to help reach a wider audience. Participants were also invited to forward the email invitation to others they thought might be interested. In addition, the link to the survey was shared on social network sites (Facebook and Twitter). Using social media as an adjunct recruitment strategy has been found to be a successful and cost-effective recruitment style [32].

When participants opened the survey link they were given information about the study and an online consenting process was activated. Once participants had given consent they could begin the online questionnaire.

Analysis

We analysed our quantitative data using simple distributions, frequencies and means using SPSS (IBM SPSS Statistics for Macintosh, Version 25.0) [33] and R (Microsoft R Open 3.5.1) [34]. All visualizations were created using Tableau Desktop and are hosted online at Tableau Public. All of the raw data is freely available and can be downloaded from https://miguems.github.io/dotheshuffle/, a webpage specifically designed to accompany this paper.

Following Braun and Clarke [35], thematic analysis was used to identify categories and themes that represented the data. Overall, categories were identified which became the first codes (level 3, L3). These codes were then grouped into themes (level 2, L2). Finally, themes were merged to develop higher-level themes (level 1, L1).

After reviewing a subset of the data (72 responses in total), two different coders (Coder 1 and Coder 2) independently identified categories to represent key elements of the data by creating level 3 (L3) codes. Both coders then compared their codes and developed larger categories (L2) to be incorporated into a codebook. (See the supplementary information (S2 File) for the full codebook with definitions and examples). A new coder (Coder 3) used the codebook to code the entire data set. To verify that the final coder had interpreted the codes as intended, the codes assigned by Coder 3 were compared to the original codes assigned by Coder 2. Coder 2 assigned more codes (L3 and L2) overall (n = 436) than Coder 3 (n = 364). 66% of Coder 3’s codes were in agreement with Coder 2’s codes. This calculation included the times when Coder 3 did not assign a code. Following this, the research team discussed any discrepancies between the codes and the final coding of all the data was then finalized. Following this, higher level themes (L1) were developed.

Data visualization

Several visualisations were created to represent the open-ended responses collected. The visualizations are available online on the project website [https://miguems.github.io/dotheshuffle/]. Developed with Tableau, they include several data filters to enable a more detailed view into distributions of the identified codes by sex and use of music within their professional lives. While this capability is not extensively used in this paper, the interactivity provided by the filters allows the exploration of relationships of the different codes with certain demographic factors and responses to the closed questions.

Survey experience

The median duration of survey completion was 17 minutes. 278 out of 322 participants (86%) stated they had not skipped any tracks (after clicking shuffle) before writing about the tracks. 12% skipped one track, either after the first or second time they shuffled their device, while only 2% skipped for both tracks. The average enjoyment level was 4.34 (SD = 0.79) out of 5 for the tracks discussed by participants that did not skip a track, while the average for the tracks discussed by those that skipped one or more tracks was 4.33 (SD = 0.66). Further investigation showed that there were no relevant differences in the distribution of assigned codes, as illustrated by the skipped track filter in the interactive visualisation. Therefore, responses by participants who had skipped tracks before responding were combined with those by participants who had not skipped tracks in the analysis. The majority (n = 387) of the 397 participants answered all the open-ended questions for two different tracks. The responses from the participants that only answered the questions for one track (n = 10) were also included in the analysis. Even though participants were not explicitly asked to do so, 83.4% (out of 322 participants who answered this question) reported listening to at least one of the tracks when writing about it. 66% (out of 322 participants who answered this question) had heard the track that they wrote about during the last month, suggesting that for the majority of the participants the tracks were relatively fresh in their minds.

For the question, “Do you have a relationship with this track? Whether it's 'Yes' or 'No' please explain”, participants responded that they did not have a relationship with a large proportion (about 50%) of the 784 total tracks discussed. Interestingly, some of those participants who answered “no” describe aspects that fit with our understanding of the term ‘relationship’ within their responses (e.g. enjoyment, used for a specific purpose, associations etc.). With some of the participants responding “no”, our findings suggest that our understanding of the word ‘relationship’ in the context of music listening may not be representative of how a substantial group of our respondents think about the music they listen to, associating this phrase with an emotional response rather than an associational one.

Results

RQ1: Why do people download and listen to music? Because they like it

The top four reasons for downloading the tracks were that the participants had liked the artist (n = 340 out of a total of 1887 options chosen, 18%), the way it sounded (n = 329, 17%), the album (n = 232, 12%) and the genre (n = 221, 12%). Choices according to basic labelling information–liking the artist, album and genre–accounted for almost half the reasons for why participants downloaded the tracks (47%). All top answers fit within our broad category of choosing to download a track because the participant liked it or something about it (Fig 1).

Fig 1. The most common reasons for downloading the tracks ordered by %.

Fig 1

The top five ranked reasons for listening to the track were “Because I really like listening to it” (n = 28, 20%), “To help carry out/enhance the activity I was doing (n = 19, 13.57%), “To accentuate an emotion/mood” (n = 13, 9,29%), “None of the above” (n = 13, 9.29%) and “To listen to the lyrics” (n = 11, 7.86%). While this data was only collected from 70 participants due to a glitch in the online survey, these top five most relevant options make up 60% of the total options ranked by those 70 participants (Fig 2). We performed a Friedman’s test to determine if the differences between rank positions were statistically significant, finding that there were significant differences in how these options were ranked χ2(12) = 167.96, p < 0.05. Post hoc analysis using Wilcoxon signed-rank tests was conducted, with a Bonferroni correction applied (p < 0.01), to test for significant differences in the top five ranked options. “Because I like listening to it” was ranked significantly higher than “To help carry out an activity” (Z = -3.566, p < 0.01). However, “To help carry out an activity” and “To accentuate a mood” were not ranked significantly differently (Z = -1.254, p = 0.21). “To accentuate a mood” was ranked significantly higher than “None of the above” (Z = -5.337, p < 0.01) that was in turn ranked significantly higher than “Listening to lyrics” (Z = -3.65, p < 0.01).

Fig 2. The first ranked reasons (ordered by % of total) chosen by the 70 participants who answered why they usually listen to the track.

Fig 2

784 tracks were included in the overall analysis. The majority of the tracks (91%, n = 719) were rated as enjoyable (enjoy (n = 336, 43%) and really enjoy (n = 383, 49%)). Few tracks (9%, n = 67) were rated by the participant as not enjoyable (1% really did not enjoy (n = 7), 2% did not enjoy (n = 14) and 6% neither enjoyed/not enjoy (n = 44)). Ten tracks were not rated for enjoyment.

RQ2: Which aspects of the music listening experience are prioritized when people talk about a track on their device? The informational context around it

Each participant shuffled their device twice and gave the title and artist of the tracks that came up. A total of 784 tracks were identified by the 397 participants who completed this portion of the questionnaire. Responses to all of these 784 tracks were included in the analysis. Of these 784 tracks some were duplicates. For example, three participants discussed Castle on the Hill by Ed Sheeran. (The track titles, artists and extracted Spotify data can be found in the supplementary material (S3 File)).

Each participant (n = 397) answered four open questions about two separate tracks: “What's the first thing that comes into your mind about this track?”, “Why did you choose to save this track to your device?”, “Why do you enjoy or not enjoy listening to the track?” (which followed the enjoyment rating) and “Do you have a relationship with this track?”. An initial reading of the responses suggested that participants often repeated information when answering the different questions and also referred back to previous answers when talking about the same track. The responses to the four questions were coded as one unit in order to avoid redundant coding within one participant’s responses about a track. The majority (97%) of participants answered these questions for two different tracks. Those that only answered the questions for one track were still included in the analysis. Each participant that answered the questions for two tracks (n = 387) had one paragraph of text for each track and each paragraph was coded separately. Differences between people who use music in their professional life and those that do not were investigated. There were no differences found in people responses to the closed questions and the themes identified from the open-ended questions.

As mentioned above, the codebook consisted of 3 levels of codes. The highest level (L1) had 4 higher-order themes. The middle level (L2) had 14 categories while the lowest level (L3) had 74 codes with a minimum of four and a maximum of 505 occurrences. Occurrence refers to the number of times that a code was discussed in relation to any track. For example, if the code “artist” was identified in one participant’s response to their first and second track this would count as two occurrences coded as artist. In total there were 4,182 code occurrences. See Fig 3 for a breakdown of the code hierarchy.

Fig 3. All codes by each level.

Fig 3

The number of occurrences that were coded for each level 3 code.

Four main higher-order (L1) themes were developed: Associations, Characteristics, Evaluations and Responses Induced (the total occurrences for each L1 theme are given in Fig 4). These themes address which aspects of the listening experience people prioritise when talking about a track on their device. Associations, as the most common type of response (with 1,946 occurrences out of the total 4,182 occurrences), were prioritised over listeners’ characterisations and evaluations of the track and how they responded to it. The themes developed from the analysis showed that participants based their evaluations, characterisations and responses not just on the music itself but also on the music-informational context around it; the artist, album, and genre.

Fig 4. The number of occurrences that were coded for each level 1 code.

Fig 4

Associations

The Associations higher order theme was used when the respondent related the track to something or someone outside of the musical track itself (e.g. "I did think of its brilliantly directed and acted music video quite a lot after watching it as it is quite a powerful video"). We identified six different L2 categories within Associations: Track Identifiers, Memory, Activity, Person, Other Media Forms, and Imagination (Fig 5 shows the breakdown of all the L2 themes). Responses were coded as Track Identifiers (the second most common code with 625 occurrences making up 15.0% of the overall codes and a third of the Associations codes (32.1%)) when the respondent associated the track with the album, artists, era, genre, or version of the track (e.g. "I have all of The Beatles' albums saved because they're obviously amazing."; “It's a fairly early Prince track from the era when he played everything himself”; “This is the O.S.T. version.”). A response was coded as Memory (531 occurrences) when the respondent associated the track with the memory of discovering the track, an event, the last time they heard it, a place, a time or stage in their life (e.g. "Remembering the last time I heard it at a party and what happened that night"; “Now every time I listen to it, it reminds me of dancing like crazy at the concert”; “If I stop to think about it I can associate events with it (playing it on the piano at home, listening to it in a concert once half a life-time ago, and imagining the performer groaning his way through some of the other tracks on the disc)”).

Fig 5. The number of occurrences that were coded for each level 2 code.

Fig 5

Participants associated their tracks with an Activity (301 occurrences; e.g. "I enjoy calm solo piano music to either read or sleep to"; “Its catchy to walk/travel to”) or a Person (206 occurrences; e.g. "Reminds me of my boyfriend when he's away from me"; “The Ozarks were a favourite group of a cousin, who I loved”). Participants also associated the tracks with Other Forms of Media (176 occurrences) for example the music video, movie, television show or musical the track appeared in (e.g. "I have been to see the musical which it is from (Phantom of the Opera) and so bought the soundtrack"; “I associate this song with the TV show called Glee.”), the playlist the track was from or the application it was found on (e.g. It might be another Discover Weekly, and I sometimes mooch around the suggestions on Spotify, or 'similar artists' links.").

Finally, another L2 category identified was Imagination (107 occurrences). A response was coded as Imagination when the respondent described the track through a descriptive narrative which is not a retelling of a memory (e.g. “The intro is very haunting so I picture a dark club with red lights that flicker through the club”; “I feel I am at a classic bar or lounge where I am smoking a Cuban cigar and sipping a drink with my friends. At the same time a group of female dancers dance according to the beat of this track”), or a description of the weather or season associated with the track (e.g. “It's quite summery and makes me think of hot summer nights”; “The first thing that comes to mind is a warm summer day and feeling the heat on my face.”; “Spinning in a field- arms out enjoying the sun”).

Characteristics

Characteristics had 1,175 occurrences (28% of the total codes) and was used when the respondent described or talked about the track by characterising it by how it sounds using expressive adjectives (e.g. lively, lovely, peaceful) or by describing the more technical aspects of the track (e.g. the beat or the lyrics). These were divided into three sub-themes: Sound Features, Lyrics, and Musical Expressive (Fig 5).

The category with the most occurrences overall and in this higher order theme was Sound/Musical features with 653 occurrences. This category makes up 15.6% of the overall codes and half (55.6%) of the codes under the Characteristics higher-order theme (Fig 3). Responses were coded as Sound/Musical Features when the respondent described the track using technical terms like beat, melody or production (e.g. “It's got a wonderfully syncopated bass line…"). With 14 different technical musical features identified, there is a wide range in the variety of the technical terms participants referred to. Some specific features identified were the production (e.g. “Troye soulfully sings the song's simple yet effective lyrics complimented by some excellent production which is paced very well feeling subtle at times and then more dramatic as the song goes on”), hook (e.g. “Melody serves as a very effective hook”) and Song Structure/Composition (e.g. “It’s one of those great tracks that builds to a climax and doesn't disappoint”) of the tracks.

Lyrics (276 occurrences) were coded separately from Sound/Musical features (e.g. beat and melody). This was because a large majority of participants talked about the lyrics specifically and did so separately to other sound features. A response was coded under Lyrics when the respondent talked about the track by mentioning or describing the lyrics or the lyrical meaning (e.g. "It has a cool meaning behind the lyrics"; “I think the lyrics are really meaningful and poetic.”). Also under the Characteristics theme was the category Musical Expressive (246 occurrences). Responses were coded as Musical Expressive when the respondent used a descriptive adjective to describe the track. The adjectives used varied widely (e.g., “Bouncy”, “Melancholy”, “Vivacious”, “Ploddy Dirgey”).

The variety in the identified musical features and the musical expressive words speaks to the variety in the music represented in this sample as well as the high level of general musical expertise and the richness of music description of music listeners today. People seem to use a variety of everyday adjectives and technical terms to talk about their music. This diversity shows the vastness in the music as well as people’s characterisations of and interactions with that music.

Evaluations

There were 685 occurrences of Evaluations. A response was coded within Evaluations when the respondent evaluated the track, an aspect of the track, their relationship to, or familiarity about the track (e.g. "No—don't like arctic monkeys"). These were divided into Positive Evaluations (481 occurrences), Familiarity (142 occurrences), and Negative Evaluations (62 occurrences) (Fig 5).

A response was coded as Positive Evaluation when the respondent made a positive evaluation of the track or an aspect of the track, for example what the song is associated with, the artist or TV show it is from (e.g. “I love musicals and Rent is spectacular, not just for the phenomenal tracks for the meaning and characters, the words, and a great sing-a-long”; “I love Beyoncé, her music makes me feel like anything is possible”; “I like the track and I really like the singers voice”) or a musical feature (e.g. “Peaceful gorgeous song, rich voice, really simple composition”; “Sassy good sassy song for the club). However, not all evaluations of people’s music on their device were positive with 9.1% (62 occurrences) being negative evaluations (e.g. “I want to skip it. When I got the track, I loved the track and now I am over it. I am just kinda over it, it is a bit annoying now”).

Participants reported different reasons for negative evaluations of tracks. One was changing taste, as seen in the previous example. Another was the fit between the music and their own state (e.g. “Sometimes I do enjoy listening, sometimes I don't—depends on my mood and the situation, as rock 'n' roll is not always what I like”). Some negative evaluations resulted from comparisons to other tracks by the same artist or within a certain album (e.g. “I like it but not as much as some of her other tracks”) and to other aspects within the same track (e.g. “Terrible chorus. Verses are decent, so is the beat during the verses but the chorus is just so boring”). Participants also gave negative evaluations of how others might judge them for having the track (e.g. “so I should be a bit ashamed of liking them. I'm not ashamed. Not much.”) and also gave clear neutral or negative judgements of the track (e.g. “This track is annoying”; “Boring. Doesn't really appeal to me”).

Responses induced

The final higher level theme, Responses Induced, was coded the least frequently with 376 codes (8.9% of the total codes) and was used when the respondent described or talked about their response to listening to track, how it made them feel or what it made them do (e.g. "Also, I feel calm when playing this track").

Under the Responses Induced higher-order theme two different categories were developed: 90.0% of Response Induced codes (338 occurrences) were coded as Changing Response, while 10.1% (38 occurrences) were coded as Matching. A response was coded as Changing Response when the respondent described or talked about how the track changes how they feel or the atmosphere (e.g. "This track boosts up my energy levels and gives me the mood to work and chill at the same time"; “Makes me feel relaxed. Evokes a calm emotional response.”; “This song makes me stand a bit on edge”). A response was coded as Matching when the respondent talked about the track matching their expectations, as useful for listening to when they wanted to match a mood, or stating that it did or did not match their current mood (e.g. "It's more upbeat than the last one, and less dirgey, more uplifting, particularly the rising harmonic progression in the middle 8, so more suited the mood I'm in which is positive right now."; “It's a good song I just am never in the mood for it”). This theme may be the most represented in previous literature but our analysis and method suggests that people do not prioritise their emotional responses to their music over their evaluations, characterizations, or associations with that music.

Results summary

Some themes in our analysis have been discussed in previous research. For example, associations with specific memories [19], activities tracks are used for [26], or expressing the different types of emotional responses tracks can induce [11]. However, other themes showed a richness previously underrepresented in the literature. For example, the many ways people associate tracks to information outside of the music itself and the variety in the way people characterise music. Moreover, the context considered when discussing the track was not just that of the music-informational context (e.g. the album or artist) but also the wider media the track may have been associated with (e.g. the music video, musical or film the track appeared in). People also associated a track with a setting and used their imaginations to create a scenario that the music represents. Finally, by using the Shuffled Play method, we were able to elicit a variety in people’s responses. This included some tracks participants did not enjoy or did not like specific aspects of. This helps highlight the nuanced way in which people talk about music.

Discussion

Over 70% of people in our study choose to download music based on what they like about it: who the artist is, the way it sounds and/or which album it is from. However, because some people also download music onto their device for other reasons–a specific event, enjoyment of an artist’s previous tracks, or using it for an activity–they do not necessarily like all their downloaded music. This highlights the importance of our method. Using Shuffled Play enables an exploration of these other types of relationships with music. Compared to previous experimental methods (i.e., researchers selecting the music themselves or asking about favourite pieces), this method is more naturalistic and similar to how people listen to music on their devices. This method also allows participants to cover a wider range of music that might be meaningful to the participant.

When considering why participants listen to a given track, comparing our results to the same question asked in Greasley and Lamont [1] we found the same top reason selected: “Because I really like listening to it”. However, none of the other top five overlapped. Perhaps the reasons lie on the different methods used: we asked participants to shuffle their music libraries whilst Greasley and Lamont [1] used experience sampling to ask people about music they heard at different moments during the day. In fact, one of the top five ranked reasons in this study was “None of the above”. By analysing responses from both open and closed questions, a better understanding of the potential reasons this option was chosen so often was achieved.

The most coded L3 code was Positive Evaluation (481 occurrences) and the top ranked reason for listening to the track was because participants liked it. Even though this may seem obvious, it is important to highlight. From this study however, it is difficult to ascertain whether “Because I like it” is an independent category or an umbrella term used that incorporates the various other more fine-grained motivations usually associated with music listening like mood regulation. It may also be through enjoyment that music can have other functions. For example, one can speculate that if people do not enjoy the music in the moment of listening, then it might not help them use music in the ways we often hear about (relax at that time, run faster, etc.). Future research could further explore the potentially moderating effect of enjoyment.

The results of this new method of data collection also highlight the importance of music-informational context as a factor when people describe their relationship with the music they have saved on their devices. Recent research has begun to highlight the relevance of music-informational context on people’s musical choices and judgements. For example, the linguistic fluency of the song title and artist [36] and the information included in program notes [37] have been found to affect people’s aesthetic judgements of music at least in some cases.

In our study we see that one important music-informational factor is the artist. This contextual information itself may override certain sound preferences: artist is the second most frequent L3 code (and is under the Track Identifier L2 theme). The results highlight the relationship that listeners have with particular artists as a driving force when describing and making judgments about music. It has previously been shown that the perceived prestige of the artists of a track can change the judgements people make about the same musical pieces when they are told they are different [21]. A performer’s status can bias the evaluation of musical stimuli [38] and empathy with an artist was found to be an important factor in aesthetic experiences of music and poetry [39]. The importance of specific people can also be seen in classical music as canonical composers have dominated music history, concert platforms and societal preferences for centuries. Our results provide a similar and supplementary result, that people prioritise the artist not just when judging but also when talking about music. By using a completely different method–primarily descriptive compared to the neurological and quantitative methods of the previous literature mentioned–we come to a similar conclusion: people’s feelings towards the artist and the informational context around the track can drive their discussion of the music, their preferences for it and whether or not to download and listen to it.

People seem to be aware of the importance of the music-informational context in their music listening and downloading experience (cf. [17]) and in our open-ended question data we find that people can be critical and acknowledge the importance of the informational context by discussing their liking of the artist or album as their reason for choosing the music. They can like the artist but not the track or they can like the track but not the artist. Very few research studies acknowledge the complex relationship between preference for an artist and preference for a particular piece of music in terms of everyday music listening experiences. It may also be the case that people have a two-step process when thinking about why they download music that might be indicative of how people search and find music online. First decisions might be made based on genre or artist preference and following this judgment, preference around the sound and the music itself may be made. This study, along with the few previous studies mentioned, exposes the need for more research investigating this important and complex relationship.

Previous research has also focused on the emotions music can evoke, arguing that this may be an important function of everyday music listening [10,11]. This study shows that this is not an aspect of the music listening experience people consistently prioritize in a Shuffled Play situation. People describe the track, characterizing what they like and do not like about it and what it reminds them of before they describe how it makes them feel. They also seem to have a wide variety of terms that they use describe the music, whether it be musical features, descriptive adjectives, or an image or scenario. Juslin’s [40] model proposes a theory that visual imagery acts as a potential mechanism for why music evokes emotion. The identification of the Imagination and Musical Expressive themes in our data set might speak to the importance of visual imagery and imagination in people’s experiences with music. People also take into consideration the lyrical content of the piece as an important element to prioritise when describing a particular piece of music. All of these findings show the range of characterizations of the music that people use when discussing music in everyday life.

While some of our results are in line with previous literature, this study uses a novel method that refocuses or reframes previous findings and thinking. This study has highlighted that enjoyment is important and should not be ignored but rather emphasised when discussing people’s relationships to music. The functions and emotions felt when listening to music can be highly affected by, and dependent on, the participants’ enjoyment as well as the informational context around the music. It is not only situational context, which has been discussed in previous literature, but also the information around the music itself that seems to be most important.

Our study offers resources to continue or extend explorations of shuffled music listening. In particular, we have made openly available the datasets with the responses, codes and demographic information, and we have created a set of interactive visualisations that can be freely accessed online and experimented with. For example, research questions about the relationship between different music genres and characteristics and the way they are described are beyond the scope of this paper but the data set lends itself to such analysis. Similarly, the visualizations available could be used to further explore the relationships between the identified themes and particular demographic factors. With our unique method, along with our interactive visualizations, we were able to explore in a novel way why people download and listen to music and which aspects of the music listening experience are prioritized when people talk about a track on their device.

Limitations

The shuffle function is used with downloaded music by some music listeners, but we do not assume that this way of music listening is representative of music listening habits across the board. We also consider that the results of this study are a reflection of the specific Shuffled Play method utilized in this study. Moreover, the shuffle function, while useful in one respect, means that only a small portion of people’s music collections were investigated. Furthermore, the shuffle function is only semi-random—the tracks that are displayed were not selected in a truly random way. Setting aside the process of the shuffle algorithm embedded in Spotify and other music listening applications, 44 participants, out of the 322 that answered this question, skipped the track which was chosen through the shuffle method for either one or both of the tracks (of these only five skipped the app-chosen track for both responses).

In addition, some applications do not have a shuffle function for a whole music library (e.g. Spotify). In this case our method limited the tracks selected to no longer represent participants’ full library but rather a selection of their library or playlist. In addition, the participants themselves, which primarily represents female, young adults, and people from or residing in the UK, cannot be assumed to generalise to all contexts and cultures. Nevertheless, despite these limitations and helped by our large sample size, the data set includes a wide variety of tracks and participant descriptions.

Due to a glitch in the online survey, only 70 participants ranked the reasons why they listen to 140 tracks. The original paper by Greasley and Lamont [1], from which this question was based on, included 25 participants. Therefore, the interpretation of these current results, though taken from a smaller subgroup of our sample, is significant in comparison with other studies in this field.

A potential area of bias in our design was in how the questions were structured in the survey. We asked questions about why they had the piece on their device that may have prompted a more practical way of discussing that music. To mitigate this, these questions were asked after all the open-ended questions had been answered. However, the task itself may have prompted more practical rather than emotional responses as we did not specifically ask them about a song that they had an emotional connection with.

Previous literature has also discussed the importance of context when discussing everyday experiences with music [1,3,6]. People chose music to put on their device for use in different contexts, which is seen in some of the responses from our participants who associate a track with a certain activity. However, the shuffled listening situation used in this study disregards this, as the track chosen is semi-random.

Implications for future research

The understanding that the music-informational context plays a significant role in people’s music listening experience should be considered when addressing other questions about musical preference, functions of music listening and musical judgements. Furthermore, research in music therapy and music in health suggests that giving people an opportunity to experience music in and of itself without a specific functional outcome is important and valuable. However, there is little research explicitly investigating where the aesthetic experience and enjoyment of music sits in relation to people’s everyday experience of music and its use in health contexts.

While this study included responses to some music that participants disliked, the participants liked the majority of tracks they discussed and no data was collected on the tracks that were skipped. A modification of this method could usefully intervene at the moment of skipping tracks to understand why people might dislike a track or in what contexts and why people might skip a track when listening to music using Shuffled Play.

Future research could also explore the musical features and their relationship to people’s descriptions of the music. We have included with this paper the extracted Spotify data (such as the track title and artist, the average bpm, and loudness) in the supporting information (S3 File). This data was extracted using Spotify’s “Sort Your Music” application (http://sortyourmusic.playlistmachinery.com/). Data like this could be used to help explore this relationship, as has been done for preferred music by Knox and MacDonald [41].

Conclusion

The novel method used in this study helps to illuminate the variety of ways people discuss and talk about music they have curated or collected on their devices. While the specific findings are not revolutionary, the new method and the mix of response type data used allows for the illumination of the diversity and subtlety of personal listening experiences. Overall, people listen to and download music because they like it. This straightforward but important observation is usually glossed over in the understanding of music listening habits, particularly when looking at the impact of functional music listening. Moreover, what people like about the music is found here to depend on a variety of important factors, mostly those outside of the musical content itself. People make critical judgements about many different aspects of the music and seem to prioritize the music-informational context along with the characterization of it over their emotional responses to it. This gives insight into what drives participants’ music preference and enjoyment. These findings may contribute to improved understanding of how music may be more appropriately used, and function better, in health, educational and social settings.

Supporting information

S1 File. Full questionnaire.

Administered online through Qualtrics.

(PDF)

S2 File. Code book.

Finalized code book used by both coders.

(PDF)

S3 File. Track and extracted Spotify data.

Track information given by the participant and track data extracted from “Sort Your Music” application.

(PDF)

Acknowledgments

We would like to thank Nordoff Robbins and the Royal College of Music for supporting the project. We would also like to thank Victoria McCrea for her help with survey design, participant recruitment and data analysis; Megan Kibler and Chloe Parks for their help with data analysis; and Li-Ching Wang and Owen Coggins for help and advice throughout.

Data Availability

All of the raw data is freely available and can be downloaded for reuse from [https://miguems.github.io/dotheshuffle/], a webpage specifically designed to accompany this paper.

Funding Statement

This work is funded by the EU H2020 programme (ga No. 743623). NS has been supported by HEartS, a project funded by the UK’s Arts and Humanities Research Council to investigate the health, economic and social impact of the arts (grant ref. AH/P005888/1) to MMS. The funder had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

References

  • 1.Greasley AE, Lamont A. Exploring engagement with music in everyday life using experience sampling methodology. Music Sci. 2011;15: 45–71. 10.1177/1029864910393417 [DOI] [Google Scholar]
  • 2.Juslin PN, Liljeström S, Västfjäll D, Barradas G, Silva A. An experience sampling study of emotional reactions to music: Listener, music, and situation. Emotion. 2008;8: 668–683. 10.1037/a0013505 [DOI] [PubMed] [Google Scholar]
  • 3.North A, Hargreaves DJ, Hargreaves JJ. Uses of music in everyday life. Music Percept. 2004;22: 41–77. [Google Scholar]
  • 4.Sloboda JA, O’Neill SA, Ivaldi A. Functions of music in everyday life: An exploratory study using the experience sampling method. Music Sci. 2001;5: 9–32. 10.1177/102986490100500102 [DOI] [Google Scholar]
  • 5.North AC, Hargreaves DJ. Musical preferences during and after relaxation and exercise. Am J Psychol. 2000;113: 43–67. 10.2307/1423460 [DOI] [PubMed] [Google Scholar]
  • 6.Krause AE, North AC, Hewitt LY. The role of location in everyday experiences of music. Psychol Pop Media Cult. 2016;5: 232–257. 10.1037/ppm0000059 [DOI] [Google Scholar]
  • 7.Greb F, Schlotz W, Steffens J. Personal and situational influences on the functions of music listening. Psychol Music. 2018;46: 763–794. 10.1177/0305735617724883 [DOI] [Google Scholar]
  • 8.Greb F, Steffens J, Schlotz W. Understanding music-selection behavior via statistical learning. Music Sci. 2018;1: 1–17. 10.1177/2059204318755950 [DOI] [Google Scholar]
  • 9.Mas-Herrero E, Marco-Pallares J, Lorenzo-Seva U, Zatorre RJ, Rodriguez-Fornells A. Individual differences in music reward experiences. Music Percept An Interdiscip J. 2013;31: 118–138. 10.1525/mp.2013.31.2.118 [DOI] [Google Scholar]
  • 10.Schäfer T, Sedlmeier P, Städtler C, Huron D. The psychological functions of music listening. Front Psychol. 2013;4 10.3389/fpsyg.2013.00004 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Baltazar M, Saarikallio S. Toward a better understanding and conceptualization of affect self-regulation through music: A critical, integrative literature review. Psychol Music. 2016;44: 1500–1521. 10.1177/0305735616663313 [DOI] [Google Scholar]
  • 12.Saarikallio S. Music as emotional self-regulation throughout adulthood. Psychol Music. 2010;39: 307–327. 10.1177/0305735610374894 [DOI] [Google Scholar]
  • 13.Van Goethem A, Sloboda JA. The functions of music for affect regulation. Lamont A, Eerola T, editors. Music Sci. 2011;15: 208–228. 10.1177/102986491101500205 [DOI] [Google Scholar]
  • 14.Ter Bogt TFM, Mulder J, Raaijmakers QAW, Nic Gabhainn S. Moved by music: A typology of music listeners. Psychol Music. 2010;39: 147–163. 10.1177/0305735610370223 [DOI] [Google Scholar]
  • 15.Saarikallio S, Maksimainen JP, Randall WM. Relaxed and connected: Insights into the emotional–motivational constituents of musical pleasure. Psychol Music. 2018; 1–19. 10.1177/0305735618778768 [DOI] [Google Scholar]
  • 16.Sallavanti MI, Szilagyi VE, Crawley EJ. The role of complexity in music uses. Psychol Music. 2016;44: 757–768. 10.1177/0305735615591843 [DOI] [Google Scholar]
  • 17.Schäfer T, Sedlmeier P. From the functions of music to music preference. Psychol Music. 2009;37: 279–300. 10.1177/0305735608097247 [DOI] [Google Scholar]
  • 18.Krause AE, Hargreaves DJ. MyTunes: Digital music library users and their self-images. Psychol Music. 2012;41: 531–544. 10.1177/0305735612440612 [DOI] [Google Scholar]
  • 19.Belfi AM, Karlan B, Tranel D. Music evokes vivid autobiographical memories. Memory. 2015;24: 979–989. 10.1080/09658211.2015.1061012 [DOI] [PubMed] [Google Scholar]
  • 20.Gabrielsson A. Strong experiences with music: Music is much more than just music (translation of Gabrielsson, A., 2008, Starka musikupplevelser—Musik är mycket mer än bara music, Hedemora: Gidlunds). Oxford, UK: Oxford University Press; 2014. [Google Scholar]
  • 21.Anglada-Tort M, Baker T, Müllensiefen D. False memories in music listening: Exploring the misinformation effect and individual difference factors in auditory memory. Memory. 2018;27: 612–627. 10.1080/09658211.2018.1545858 [DOI] [PubMed] [Google Scholar]
  • 22.Greasley AE, Lamont A, Sloboda J. Exploring musical preferences: An in-depth qualitative study of adults’ liking for music in their personal collections. Qual Res Psychol. 2013;10: 402–427. 10.1080/14780887.2011.647259 [DOI] [Google Scholar]
  • 23.Krause AE, North AC, Hewitt LY. Music-listening in everyday life: Devices and choice. Psychol Music. 2015;43: 155–170. 10.1177/0305735613496860 [DOI] [Google Scholar]
  • 24.Krause AE, North AC. Contextualized music listening: Playlists and the Mehrabian and Russell model. Psychol Well-Being Theory, Res Pract. 2014;4: 1–16. 10.1186/s13612-014-0022-7 [DOI] [Google Scholar]
  • 25.Avdeeff M. Technological engagement and musical eclecticism: An examination of contemporary listening practices. J Audience Recept Stud. 2012;9: 265–285. [Google Scholar]
  • 26.Heye A, Lamont A. Mobile listening situations in everyday life: The use of MP3 players while travelling. Music Sci. 2010;14: 95–120. 10.1177/102986491001400104 [DOI] [Google Scholar]
  • 27.Müllensiefen D, Gingras B, Musil J, Stewart L. The musicality of non-musicians: An index for assessing musical sophistication in the general population. Snyder J, editor. PLoS One. 2014;9: e89642 10.1371/journal.pone.0089642 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.Müllensiefen D, Gingras B, Musil J, Stewart L. The musicality of non-musicians: An index for assessing musical sophistication in the general population. Snyder J, editor. PLoS One. 2014;9: e89642 10.1371/journal.pone.0089642 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29.Qualtrics. Qualtrics. Utah, USA: Qualtrics; 2019. Available: https://www.qualtrics.com
  • 30.Coleman J. Relational analysis: The study of social organizations with survey methods. Hum Organ. 1958;17: 28–36. 10.17730/humo.17.4.q5604m676260q8n7 [DOI] [Google Scholar]
  • 31.Goodman LA. Snowball sampling. Ann Math Stat. 1961;32: 148–170. 10.1214/aoms/1177705148 [DOI] [Google Scholar]
  • 32.Khatri C, Chapman SJ, Glasbey J, Kelly M, Nepogodiev D, Bhangu A, et al. Social media and internet driven study recruitment: Evaluating a new model for promoting collaborator engagement and participation. PLoS One. 2015;10: e0118899 10.1371/journal.pone.0118899 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.BM Corp. BM SPSS Statistics for Windows, Version 25.0. Armonk, NY: IBM Corp; 2017. [Google Scholar]
  • 34.R Core Team. R: A language and environment for statistical computing. Vienna, Austria: R Foundation for Statistical Computing; 2016. [Google Scholar]
  • 35.Braun V, Clarke V. Using thematic analysis in psychology. Qual Res Psychol. 2006;3: 77–101. 10.1191/1478088706qp063oa [DOI] [Google Scholar]
  • 36.Anglada-Tort M, Steffens J, Müllensiefen D. Names and titles Matter: The impact of linguistic fluency and the affect heuristic on aesthetic and value judgements of music. Psychol Aesthetics, Creat Arts. 2018. [cited 22 Feb 2019]. 10.1037/aca0000172 [DOI] [Google Scholar]
  • 37.Margulis EH. When program notes don’t help: Music descriptions and enjoyment. Psychol Music. 2010;38: 285–302. 10.1177/0305735609351921 [DOI] [Google Scholar]
  • 38.Aydogan G, Flaig N, Ravi SN, Large EW, McClure SM, Margulis EH. Overcoming bias: Cognitive control reduces susceptibility to framing effects in evaluating musical performance. Sci Rep. 2018;8: 6229 10.1038/s41598-018-24528-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 39.Margulis EH, Levine WH, Simchy-Gross R, Kroger C. Expressive intent, ambiguity, and aesthetic experiences of music and poetry. Jaencke L, editor. PLoS One. 2017;12: e0179145 10.1371/journal.pone.0179145 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 40.Juslin PN. From everyday emotions to aesthetic emotions: Towards a unified theory of musical emotions. Phys Life Rev. 2013;10: 235–266. 10.1016/j.plrev.2013.05.008 [DOI] [PubMed] [Google Scholar]
  • 41.Knox D, MacDonald R. Broadcasting personalities: The relationship between occupation and music preferences in the BBC Radio programme Desert Island Discs. Psychol Music. 2017;45: 645–664. 10.1177/0305735616670497 [DOI] [Google Scholar]

Decision Letter 0

Sarah EP Munce

14 Oct 2019

PONE-D-19-17641

Do the Shuffle: Using shuffled play to explore reasons for music listening

PLOS ONE

Dear Neta Spiro,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

In particular, the mixed methods design of your study needs to be specified and integration needs to be addressed.  It is recommended that you integrate your quantitative and qualitative findings narratively and via a joint display.  Furthermore, the novel contribution of your work should be elucidated.

We would appreciate receiving your revised manuscript by 11 January 2020. When you are ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter.

To enhance the reproducibility of your results, we recommend that if applicable you deposit your laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). This letter should be uploaded as separate file and labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. This file should be uploaded as separate file and labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. This file should be uploaded as separate file and labeled 'Manuscript'.

Please note while forming your response, if your article is accepted, you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out.

We look forward to receiving your revised manuscript.

Kind regards,

Sarah E.P. Munce, PhD

Academic Editor

PLOS ONE

Journal Requirements:

1. When submitting your revision, we need you to address these additional requirements.

Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at

http://www.journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and http://www.journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

Additional Editor Comments (if provided):

Thank you for your submission to PLOS ONE. Please review and address all of the comments of the individual reviewers. In particular, the authors need to specify the mixed methods design of this study and integrate the findings of the qualitative and quantitative components (narratively and via joint display). The authors should also clearly elucidate the novel contribution of this work.

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Yes

Reviewer #3: No

Reviewer #4: Partly

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: Yes

Reviewer #3: Yes

Reviewer #4: Yes

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

Reviewer #3: Yes

Reviewer #4: Yes

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

Reviewer #3: Yes

Reviewer #4: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: In this paper the authors examined factors that lead individuals to download and/or listen to specific music selections. Rather than responding to pre-selected music examples or selecting their own music examples, participants responded to two pieces that were generated by their device’s “shuffle” feature. This was an innovative design twist that allowed for a potentially richer set of data, tapping into choices that at once represented the participants’ own volition yet did not necessarily represent their strongest preferences.

Overall, I found this to be an interesting paper that nicely balanced precision necessary for survey inquiry with the more natural context-dependent reality of personal music choices. As suggested by the authors, the findings were not ground-breaking in their novelty; rather, they underscored the fluidity and diversity of the personal music listening experience.

I do have several comments and suggestions for the authors. These are largely focused on aspects needing greater clarity or consistency, however, as such, are consequential for helping the reader to understand and interpret the data and interpretation presented.

1. Abstract - The title/abstract makes reference to both exploring “reasons for music listening” and the innovative aspect of the shuffle play method. Each of these issues receives different emphases throughout the paper making it somewhat unclear exactly which is the main focus of the paper. Was the main takeaway the insight into listening or the viability of the method?

2. Abstract – To what degree might “because they like it” subsume the various other more fine-grained motivations usually associated with music listening (e.g., mood regulation)? Is “liking” really a different category of reason or a broader umbrella term used by the participants to describe a range of listening experiences?

3. P. 3, line 46 – “of the time” seems a bit vague. Does this refer to any time? Some specific subset of time?

4. P. 5, line 112 – This is unclear. Was shuffle or travelling the largest category? Category of what?

5. P. 6, line 121 – Here and throughout, I am not sure this represents a “mixed methods” approach as formally defined by sources such as Creswell (2010). The use of both closed- and open-ended questions provided two different types of data, however both were the result of a large-scale survey method (and, in fact, both gathered from the same group of people, using the same instrument, at the same time). Consider using a different term to convey use of this mixed response type.

6. Table 1 – This breakdown of the sample is confusing as the different subsamples have not yet been explained in the report. The note is only marginally helpful because it references a survey item not yet presented in the method.

7. Table 1 - The diverse array of countries represented is impressive, however the vast majority of the participants are associated with the UK. This should be discussed in the limitations section; the potential impact of this imbalance is worth some speculative commentary in the discussion and should be taken into account when interpreting the results.

8. Tables 1 & 2 – For the data presented in tables 1 and 2, is there a particular need to disaggregate the sample into two groups? Differences between the two groups are not discussed among the results nor are potential relationships between the n=70 subsample and their response. Thus, the two groups may not warrant detail here.

9. Throughout much of the second portion of the paper the language was quite confusing between the responses participants provided and the participants themselves. For instance, if I understand the method correctly (p. 11, lines 169-170), each participant generated two shuffled tracks and provided a different set of responses for each.

a. On this point, did the participants complete a full questionnaire for one track and then begin anew for the second track? Or did they generate the two tracks and then provide two responses to each survey item?

b. P. 14, line 252 – Beginning around this point, the data refer to “participants” (for example, here it reads “50% of people responded negatively”). But the unit of analysis is responses to tracks, correct? Presuming that 50% of people did not respond negatively to both tracks, how many negative responses were there?

c. P. 15, line 296 - Again confusing. Each participant answered four open questions twice!

d. P. 27, line 561 – “13.6% of participants”…again confusing in its reference to participants rather than responses.

e. Overall, please reconcile language like this throughout.

10. P. 13 – The section on coding and analysis was confusing.

a. Line 222 - Were codes chosen from among those created? Were the initial codes created before this step? What do you mean by “initial review of the data”?

b. Line 227- “had more codes”; Do you mean “assigned more codes”? Do these refer to the L3 codes?

c. Line 228 - Do you mean that 66% of Coder 3’s codes were in agreement with those of Coder 2? Does this account for missing codes? In other words, did the two coders assign different codes in the remaining 34% of the cases IN ADDITION TO the number of cases in which Coder 3 did not assign a code. The level of agreement is really rather low as it is, and would be considerably (arguably unacceptably) lower if “missing” codes are over and above this figure.

d. Line 230 – “memos were written to explicate”; I understand the reference to memos, but how were they used to accomplish this purpose?

11. P. 14, Lines 243-244 - This is confusing. Each participant wrote about two tracks, correct? Following from this, what % of the track-skippers were repeat offenders?

12. P. 14, Line 249 - Again, these numbers seem to be referring to individuals when each set of track data represents 1/2 of each participant’s contribution. (Also, please indicate if and how often you included responses from participants who only evaluated one track.)

13. P. 15, line 270 -- It would be useful to remind readers here that these data only refer to the responses of 70 participants.

14. P. 24, line 471 -- It should be pointed out that this finding emerged from a smaller subset of participants. Given this significant limitation, I am wondering whether this finding should occupy such a prominent placement among the results.

15. There are several uses of the term “methodology” throughout the discussion. I think of this term as referencing the collective set of methods used in a given discipline or research tradition. Within the scope of a single study, might the more basic “method” be more accurate?

16. P. 26, line 521-522 -- Might this be a reflection of the shuffle aspect? This may be particularly so in light of self-selection as the top listening approach.

17. P. 27, lines 563-564 -- This might be worth some extra emphasis given the high percentage of participants who used Spotify.

18. Figure 2 -- Does the rank of this response call into question the completeness of the item? Were one or more vital responses left out? Do your data provide evidence of what these might include?

Reviewer #2: This is a nicely done study which builds on the work of others (particularly the fourth author) and makes a modest but worthwhile contribution to the literature. The "shuffle play" method itself is interesting and worthwhile, but in connection with the results reported herein shows a limitation to the study: the question of "why did I download this track?" gets answered in terms of artist/genre, but clearly that broad determination does not indicate how much a track would be preferred or meaningful to the listener; there's a two-stage preference sort here, with artist/genre as a first, quite rough cut, and actual listening choice/preference/response a the second, more fine-grained filter. This points out the relatively indiscriminate nature of seeking out and using online information of any kind, not limited to music.

In any qualitative study it's hard to know how to present the results to be both brief enough for publication and deep enough to be informative and to dispel questions of bias, intentional or unintentional, in how the results are being reported. I'd personally like more examples of responses, even if only in the most oft-used categories (musical characteristics, change of feelings, etc).

It would also be really interesting to get more on the _negative_ responses and, if possible, on what tracks were skipped! This method could be quire illuminating relative to the normal "preferred" and "self-chosen" music methods but the 'contrary' side--pieces which were disliked or skipped--needs to be fleshed out more for this to be the case.

Reviewer #3: I agree that this is a novel method that provides some new insights -- in particular, insights regarding the importance of associations of tracks to information outside of the music itself. See however comments below regarding a potential distortion with regard to intrretation of this finding

Regarding the reporting of data -- I found it curious that the authors did not take advantage of the opportunity to describe differences between musicians and non-musicians -- this data appears to be available

Lines 235-236. Using Tableau, I wasn't sure whether the modifications I was making to the filters were having an impact on the visualized data because of the proportion of each subpopulation in the total sample or whether it reflected the unique views of the population remaining after filtering. I wonder if using proportions would be more informative.

Lines 254-255, Possible missing details -- change to? "some of the participants who responded NO then went on to describe ..."

Lines 313-315. for me at least, "relationship with a track" does not connote anything about emotion, regardless of what the piece is -- this is about my history with the band, song, or my associations with it -- hence, I don't see this contradicting the quantitative ranked data regarding "why I choose to listen". This interpretation seems like a bit of a distortion. I believe this point appears in different places in the manuscript --- abstract and discussion -- should be reconciled throughout

Lines 455-458. the authors note that the ability to get insights about music that listeners do not like is a unique opportunity afforded by the shuffle methodology. but this is hardly the most effective means of getting this info - - they could have asked people to name and defend tracks they'll especially dislike. The authors should consider emphasizing that this is more of a "fringe" benefit of the method rather than something it is ideally designed to address

Reviewer #4: "Do the Shuffle: Using shuffled play to explore reasons for music listening" is an exciting title that piqued my curiosity. A laudable contribution to research in music on why people download and listen to music, and which aspects of music listening experiences are people prioritize while listening to music on their devices.

I am surprised to see the names of the authors on the page, making me question the blindness of this review.

The abstract is clear and answers the purpose of the study. I am curious about the tact that people download music because they like it – Seems to highlight the emotion of liking. Earlier, the authors state that participants use "sound and musical features" to characterize music and associate the music with their context "over ... emotional responses". "Therefore, how their prioritize music seems to contradict why the music is downloaded in the first place — an interesting dilemma.

The introduction: A well-posited argument, the authors emphasize the importance of self-chosen music for research, music fit and context, functions of music listening, and music preference. They address the gaps in previous studies – music is preselected for participants — the importance of self-chosen music in the new technological era. Arguing that listeners today "carry their music with them," the authors use Shuffle Play an app that allows for the flexibility of listening choices, to collect data. The App is distinct because it places ownership in the hands of the listener while also allowing for a chance. In their introduction, the authors defend their arguments answering why they chose to research self-directed listening in the first place.

Aims and Questions: Aim to use a "mixed methods approach" to answering two questions: first, why people download and listen to music, and second, which aspects of the music listening experiences are prioritized by people listening to music on their devices.

Methods: Design: Even though the researchers tell us that they are using a mixed-methods approach, we do not know what kind of mixed methods approach. In mixed-methods, researchers collect both quantitative and qualitative data. Therefore, the plan may have combined methods, but the design is either, for example, Explanatory, sequential, convergent, or complex with embedded core designs. I encourage the researchers to read: Creswell, J. W., & Creswell, J. D. (2018). Research design: Qualitative, quantitative, and mixed methods approaches (5th ed.). Thousand Oaks, CA: Sage. They are using only a survey questionnaire for this purely descriptive study. Are the authors suggesting that Shuffle play is the qualitative part of their data collection?

Participants: The authors have a robust sample size (n.322). Excellent. It will be interesting to find out why 2/3rds are female, not sure how the authors will discuss this later. Interesting that they came from different countries. How was the sample accessed? Were they invited to participate? From a conference. Not sure about the representative sample argument – if we have single-digit representations from some countries. Using snowballing techniques from authors networks (See Procedure) does not make this representative.

Shuffle play: This has loads of potential for future research. I am not sure if I can call this qualitative data collection, primarily as a questionnaire, is also used to ask descriptive questions.

Results: Associations, as the most influential type of response then, describing characteristics of a soundtrack and finally evaluation of the participant relationship, or familiarity with the music. How participants responded to music-related directly to their emotions – that is, they liked it.

My final thoughts: While this is a very promising study, it might be best for the authors to position it as a descriptive study using two forms of data collections- one through direct survey and other through online survey within a listening app. More work is needed to support the explanation of your methods.

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: Yes: Steven J. Morrison

Reviewer #2: No

Reviewer #3: No

Reviewer #4: No

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files to be viewed.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email us at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2020 Feb 6;15(2):e0228457. doi: 10.1371/journal.pone.0228457.r002

Author response to Decision Letter 0


10 Dec 2019

We would like to thank all reviewers for their thoughtful and helpful comments. Below we detail our responses to each reviewers’ comments (in italics) as well as highlight any substantial changes to the manuscript, both by including new text in this document (with page and line numbers) and providing its position in our revised document (attached).

Editor:

In particular, the authors need to specify the mixed methods design of this study and integrate the findings of the qualitative and quantitative components (narratively and via joint display). The authors should also clearly elucidate the novel contribution of this work.

Thank you for your comments. We have changed the description of the method, as per the suggestions made by the reviewers, to be a descriptive study using mixed response type as opposed to a mixed methods approach. This has been changed throughout the manuscript. Changes in the document have been made to reflect this:

With this method, along with collecting mixed response type data, we build on and go beyond the more typical foci in the literature on music and emotion, music in everyday life and music and technology

(Aims and research questions: Page 7, line 140-141)

For this descriptive study, we collect responses from a mix of question response types, including closed and open questions.

(Design: Page 7, line 152-153)

By analysing responses from both open and closed questions a better understanding of the potential reasons this option was chosen so often can be achieved.

(Discussion: Page 26, lines 524-526)

In terms of the integration of the different types of responses, we have added filters within the visualisations that enable the exploration of how these types of responses relate. We have also pointed to this within the manuscript.

While this capability is not extensively used in this paper, the interactivity provided by the filters allows the exploration of relationships of the different codes with certain demographic factors and responses to the closed questions.

(Data Visualisations, Page 15, line 267)

We have highlighted the main contribution of the study within the conclusion of the manuscript.

While the specific findings are not revolutionary, the new method and the mix of response type data used allows for the illumination of the diversity and subtlety of personal listening experiences.

(Conclusion, Page 32, lines 664-666)

Reviewer 1:

In this paper the authors examined factors that lead individuals to download and/or listen to specific music selections. Rather than responding to pre-selected music examples or selecting their own music examples, participants responded to two pieces that were generated by their device’s “shuffle” feature. This was an innovative design twist that allowed for a potentially richer set of data, tapping into choices that at once represented the participants’ own volition yet did not necessarily represent their strongest preferences.

Overall, I found this to be an interesting paper that nicely balanced precision necessary for survey inquiry with the more natural context-dependent reality of personal music choices. As suggested by the authors, the findings were not ground-breaking in their novelty; rather, they underscored the fluidity and diversity of the personal music listening experience.

Thank you for your kind comments

I do have several comments and suggestions for the authors. These are largely focused on aspects needing greater clarity or consistency, however, as such, are consequential for helping the reader to understand and interpret the data and interpretation presented.

1. Abstract - The title/abstract makes reference to both exploring “reasons for music listening” and the innovative aspect of the shuffle play method. Each of these issues receives different emphases throughout the paper making it somewhat unclear exactly which is the main focus of the paper. Was the main takeaway the insight into listening or the viability of the method?

From our perspective the main aim is to gain insight into music listening through this new method. As it is a new method we wanted to make sure we explained it clearly throughout the paper. While keeping the Do the Shuffle short title, we have changed the longer title to now mention the aim of investigating music listening first then the shuffled play method.

Do the Shuffle: Exploring reasons for music listening though shuffled play

(Title: Page 2, lines 24-25)

2. Abstract – To what degree might “because they like it” subsume the various other more fine-grained motivations usually associated with music listening (e.g., mood regulation)? Is “liking” really a different category of reason or a broader umbrella term used by the participants to describe a range of listening experiences?

We agree that this is something that needs to be more carefully discussed the data collected here does not help distinguish whether this two-step process is being used. We have added a sentence in the discussion to acknowledge this.

From this study however, it is difficult to ascertain whether “Because I like it” is an independent category or an umbrella term used that incorporates the various other more fine-grained motivations usually associated with music listening like mood regulation.

(Discussion: Page 26, lines 529-532)

3. P. 3, line 46 – “of the time” seems a bit vague. Does this refer to any time? Some specific subset of time?

This refers to participants’ waking hours, as tested through a range of experience sampling studies. The phrase has been replaced with ‘of people’s waking hours’. (Introduction, Page 4, line 65)

4. P. 5, line 112 – This is unclear. Was shuffle or travelling the largest category? Category of what?

All the participants in this study were travelling, hence ‘mobile participants’. This has been rephrased to the following to clarify the shuffle category point:

the mobile participants in Heye and Lamont's (26) study were listening via shuffle mode (the largest response category) while travelling.

(Introduction, Page 6, line 131)

5. P. 6, line 121 – Here and throughout, I am not sure this represents a “mixed methods” approach as formally defined by sources such as Creswell (2010). The use of both closed- and open-ended questions provided two different types of data, however both were the result of a large-scale survey method (and, in fact, both gathered from the same group of people, using the same instrument, at the same time). Consider using a different term to convey use of this mixed response type.

After looking at the reference you provided we agree and have changed our method to reflect the mixed response type used in this study as opposed to a mixed methods approach. This has been changed throughout the manuscript.

With this method, along with collecting mixed response type data, we build on and go beyond the more typical foci in the literature on music and emotion, music in everyday life and music and technology

(Aims and research questions: Page 7, lines 140-141)

For this descriptive study, we collect responses from a mix of question response types, including closed and open questions.

(Design: Page 7, lines 152-153)

By analysing responses from both open and closed questions a better understanding of the potential reasons this option was chosen so often can be achieved.

(Discussion: Page 26, lines 524-526)

6. Table 1 – This breakdown of the sample is confusing as the different subsamples have not yet been explained in the report. The note is only marginally helpful because it references a survey item not yet presented in the method.

We originally included this to be as transparent as possible but see that it is more confusing then helpful, especially as there were no significant differences in the demographic information between the subsamples. Therefore, we have taken these subsamples out of Table 1 and 2 and included a note below Table 1.

Note. Only 70 out of the 322 answered the question “Why do you usually listen to this track?” (described below) due to a glitch in the survey program. No significant differences in demographic information or musical experience (Table 2) were found between those that answered this question and those that did not.

(Table 1: Page 10, lines 167-170)

7. Table 1 - The diverse array of countries represented is impressive, however the vast majority of the participants are associated with the UK. This should be discussed in the limitations section; the potential impact of this imbalance is worth some speculative commentary in the discussion and should be taken into account when interpreting the results.

We recognize that this is a limitation and have made this clearer within the limitations section.

In addition, the participants themselves, which primarily represents female, young adults, and people from or residing in the UK, cannot be assumed to generalise to all contexts and cultures

(Limitations: Page 30, lines 621-623)

8. Tables 1 & 2 – For the data presented in tables 1 and 2, is there a particular need to disaggregate the sample into two groups? Differences between the two groups are not discussed among the results nor are potential relationships between the n=70 subsample and their response. Thus, the two groups may not warrant detail here.

Thanks again and we have followed your helpful suggestion. We have now taken out the subgroups in the tables and included this note for Table 2:

Note. Only 70 out of the 322 answered the question “Why do you usually listen to this track?” (described below) due to a glitch in the survey program. No significant differences in demographic information (Table 1) or musical experience were found between those that answered this question and those that did not.

(Table 1: Page 10 , lines 167-170)

9. Throughout much of the second portion of the paper the language was quite confusing between the responses participants provided and the participants themselves. For instance, if I understand the method correctly (p. 11, lines 169-170), each participant generated two shuffled tracks and provided a different set of responses for each.

a. On this point, did the participants complete a full questionnaire for one track and then begin anew for the second track? Or did they generate the two tracks and then provide two responses to each survey item?

We have made revisions to try and clarify our methods and interpretation making sure to make clear when we are talking about the participants versus the responses the participants gave to two different tracks.

Once they had responded to all open-ended questions for both tracks they went onto the second section of the questionnaire.

(A New Data Collection Method: Page 12, lines 194-195)

Participants were presented with all the closed questions to be answered in relation to the first track they discussed in the previous section. They were reminded of the track title and artist in the questionnaire. They then answered all the questions about the second track they had discussed in the previous section, again being reminded of the track’s title and artist.

(A New Data Collection Method: Page 12, lines 198-202)

An initial reading of the responses suggested that participants often repeated information when answering the different questions and also referred back to previous answers when talking about the same track. The responses to the four questions were coded as one unit in order to avoid redundant coding within one participant’s responses about the same track. The majority (97%) of participants answered these questions for two different tracks. Those that only answered the questions for one track were still included in the analysis. Each participant that answered the questions for two tracks (n = 387) had one paragraph of text for each track and each paragraph was coded separately. Differences between people who use music in their professional life and those that do not were investigated. There were no differences found in people responses to the closed questions and the themes identified from the open-ended questions.

(RQ2: Page 18, lines 342-350)

Occurrence refers to the number of times that a code was discussed in relation to any track. For example, if the code “artist” was identified in one participant’s response to their first and second track this would count as two occurrences coded as artist. In total there were 4,182 code occurrences. See Fig 3 for a breakdown of the code hierarchy.

(RQ2: Page 19, lines 355-357)

b. P. 14, line 252 – Beginning around this point, the data refer to “participants” (for example, here it reads “50% of people responded negatively”). But the unit of analysis is responses to tracks, correct? Presuming that 50% of people did not respond negatively to both tracks, how many negative responses were there?

We have changed this to now be clearer.

For the question, “Do you have a relationship with this track? Whether it's 'Yes' or 'No' please explain”, participants responded that they did not have a relationship with a large proportion (about 50%) of the 784 total tracks discussed.

(Survey Experience: Page 16, lines 286-288)

c. P. 15, line 296 - Again confusing. Each participant answered four open questions twice!

Thanks again. This has been made clearer in our wording to reflect that the unit of analysis is track rather than participant.

784 tracks were included in the overall analysis. The majority of the tracks (91%, n = 719) were rated as enjoyable (enjoyed (n =336, 43%) or really enjoyable (n =383, 49%)). Few tracks (9%, n = 67) were not rated by the participant as enjoyable (1% really did not enjoy (n = 7), 2 % did not enjoy (n =14) and 6% neither enjoyed/not enjoy (n =44). Ten tracks were not rated for enjoyment.

(RQ1: Page 17, lines 322-326)

d. P. 27, line 561 – “13.6% of participants”…again confusing in its reference to participants rather than responses.

This was made clearer with the numbers made more explicit.

Setting aside the process of the shuffle algorithm embedded in Spotify and other music listening applications, 44 participants, out of the 322 that answered this question, skipped the track which was chosen through the shuffle method for either one or both of the tracks (of these only five skipped the app-chosen track for both responses).

(Limitations, Page 30, lines 614-618)

e. Overall, please reconcile language like this throughout.

This has now been amended throughout.

10. P. 13 – The section on coding and analysis was confusing.

a. Line 222 - Were codes chosen from among those created? Were the initial codes created before this step? What do you mean by “initial review of the data”?

b. Line 227- “had more codes”; Do you mean “assigned more codes”? Do these refer to the L3 codes?

c. Line 228 - Do you mean that 66% of Coder 3’s codes were in agreement with those of Coder 2? Does this account for missing codes? In other words, did the two coders assign different codes in the remaining 34% of the cases IN ADDITION TO the number of cases in which Coder 3 did not assign a code. The level of agreement is really rather low as it is, and would be considerably (arguably unacceptably) lower if “missing” codes are over and above this figure.

d. Line 230 – “memos were written to explicate”; I understand the reference to memos, but how were they used to accomplish this purpose?

We have now rewritten this whole section to make our analysis clearer. Overall, we made the language more succinct and free of unnecessary jargon. We have also addressed all the suggestions made in a-d within these changes highlighted below.

Following Braun and Clarke (34), thematic analysis (34) was used to identify categories and themes that represented the data. Overall, categories were identified which became the first codes (level 3, L3). These codes were then grouped into themes (level 2, L2). Finally, themes were merged to develop higher-level themes (level 1, L1).

After reviewing a subset of the data (72 responses in total), two different coders (Coder 1 and Coder 2) independently identified categories to represent key elements of the data by creating level 3 (L3) codes. Both coders then compared their codes and developed larger themes (L2) to be developed into a codebook. (See the supplementary information (S2 File) for the full codebook with definitions and examples). A new coder (Coder 3) used the codebook to code the entire data set. To verify that the final coder had interpreted the codes as intended, the codes assigned by Coder 3 were compared to the original codes assigned by Coder 2. Coder 2 assigned more codes (L3 and L2) overall (n = 436) than Coder 3 (n = 364). 66% of Coder 3’s codes were in agreement with Coder 2’s codes. This calculation included the times when Coder 3 did not assign a code. Following this, the research team discussed any discrepancies between the codes and the final coding of all the data was then finalized. Following this, higher-level themes (L1) were developed in preparation for writing this paper.

(Analysis: Page 14, lines 244-259)

11. P. 14, Lines 243-244 - This is confusing. Each participant wrote about two tracks, correct? Following from this, what % of the track-skippers were repeat offenders?

We have now made this clearer by clearly stating the % of participants who skipped either one or both.

278 out of 322 participants (86%) stated they had not skipped any tracks (after clicking shuffle) before writing about the tracks. 12% skipped one track, either after the first or second time they shuffled their device, while only 2% skipped for both tracks.

(Survey Experience, Page 15, lines 269-272)

12. P. 14, Line 249 - Again, these numbers seem to be referring to individuals when each set of track data represents 1/2 of each participant’s contribution. (Also, please indicate if and how often you included responses from participants who only evaluated one track.)

We have made this clearer in this section by stating that the enjoyment levels were calculated by track not participant. We also explained in more detail how many participants answered the open-ended questions about two tracks as opposed to one.

The average enjoyment level was 4.34 (0.79) out of 5 for the tracks discussed by participants that did not skip a track, while the average for the tracks discussed by those that skipped one or more tracks was 4.33 (0.66). Further investigation showed that there were no relevant differences in the distribution of assigned codes, as illustrated by the skipped track filter in the interactive visualisation. Therefore, responses by participants who had skipped tracks before responding were combined with those by participants who had not skipped tracks in the analysis. The majority (n = 387) of 397 participants answered all the open-ended questions for two different tracks. The responses from the participants that only answered the questions for one track (n = 10) were also included in the analysis.

(Survey Experience, Page 15, lines 272-281)

13. P. 15, line 270 -- It would be useful to remind readers here that these data only refer to the responses of 70 participants.

This is a helpful suggestion and a small clarification has been included.

While this data was only collected from 70 participants due to a glitch in the online survey, these top five most relevant options make up 60% of the total options ranked by those 70 participants (Fig 2).

(RQ1: Page 17, lines 308-310)

14. P. 24, line 471 -- It should be pointed out that this finding emerged from a smaller subset of participants. Given this significant limitation, I am wondering whether this finding should occupy such a prominent placement among the results.

We have pointed this out now within the results (above) and we acknowledge this is a limitation in the limitations section. Having said that, the original paper on which the question was based included only 25 participants. We therefore feel that though this is a small subset of participants compared to the rest of the study, it is still a useful sample and can provide a contribution to our understanding of reasons why people listen to music.

Due to a glitch in the online survey, only 70 participants ranked the reasons why they listen to 140 tracks. The original paper by Greasley and Lamont (1), from which this question was based on, included 25 participants. Therefore, the interpretation of these current results, though taken from a smaller subgroup of our sample, is significant in comparison with other studies in this field

(Limitations: Page 30, lines 627-631)

15. There are several uses of the term “methodology” throughout the discussion. I think of this term as referencing the collective set of methods used in a given discipline or research tradition. Within the scope of a single study, might the more basic “method” be more accurate?

Methodology has now been changed to method throughout the discussion.

16. P. 26, line 521-522 -- Might this be a reflection of the shuffle aspect? This may be particularly so in light of self-selection as the top listening approach.

We have made this clear within the section by including shuffled play as an important aspect which contributes to the specific results interpreted within the study.

This study shows that this is not an aspect of the music listening experience that people consistently prioritize in a shuffled play situation.

(Discussion, Page 28, lines 576-578)

We have also added a sentence in the limitation section to make it clear we do understand the impact shuffled listening has on people’s responses.

We also consider that the results of this study are a reflection of the specific Shuffled Play method utilized in this study.

(Limitations: Page 30, lines 610-611)

17. P. 27, lines 563-564 -- This might be worth some extra emphasis given the high percentage of participants who used Spotify.

We have added another sentence to further explain how this limitation affects the interpretation of the results.

In this case our method limited the tracks selected to no longer represent participants’ full library but rather a selection of their library or playlist.

(Limitations, Page 30, lines 620-621)

18. Figure 2 -- Does the rank of this response call into question the completeness of the item? Were one or more vital responses left out? Do your data provide evidence of what these might include?

This is an interesting point but we have no way of knowing what responses might have been left out. Our rational for doing a top five ranking was to force people to think about the responses which represented the most salient or important reasons to help us hone in on the most important possibilities.

Reviewer #2:

This is a nicely done study which builds on the work of others (particularly the fourth author) and makes a modest but worthwhile contribution to the literature. The "shuffle play" method itself is interesting and worthwhile, but in connection with the results reported herein shows a limitation to the study: the question of "why did I download this track?" gets answered in terms of artist/genre, but clearly that broad determination does not indicate how much a track would be preferred or meaningful to the listener; there's a two-stage preference sort here, with artist/genre as a first, quite rough cut, and actual listening choice/preference/response a the second, more fine-grained filter. This points out the relatively indiscriminate nature of seeking out and using online information of any kind, not limited to music.

Thank you for this interesting point. We can understand how that can be one interpretation of the results and a potential limitation of our method. We cannot say, based on our collected data, if this filtering is being done by our participants or if they are even cognizant of making these types of stepped decisions. We do however think this is an important possibility to consider and therefore have discussed this further in the discussion section of the paper.

It may also be the case that people have a two-step process when thinking about why they download music that might be indicative of how people search and find music online. First decisions might be made based on genre or artist preference and following this judgment, preference around the sound and the music itself may be made.

(Discussion, Page 28, lines 568-572)

In any qualitative study it's hard to know how to present the results to be both brief enough for publication and deep enough to be informative and to dispel questions of bias, intentional or unintentional, in how the results are being reported. I'd personally like more examples of responses, even if only in the most oft-used categories (musical characteristics, change of feelings, etc).

We will include the full data set with the responses and assigned codes in the supplementary material. We also have added a tab to the visualization called “Participants responses" which includes the examples of responses.

It would also be really interesting to get more on the _negative_ responses and, if possible, on what tracks were skipped! This method could be quire illuminating relative to the normal "preferred" and "self-chosen" music methods but the 'contrary' side--pieces which were disliked or skipped--needs to be fleshed out more for this to be the case.

We agree that this would be interesting but unfortunately we did not collect this information. We discuss what we can about the negative responses to the tracks but few tracks were disliked. We have added sentences in the future directions section about the possibility to study this in future research.

While this study included responses to some music that participants disliked, the participants liked the majority of tracks they discussed and no data was collected on the tracks that were skipped. A modification of this method could usefully intervene at the moment of skipping tracks to understand why people might dislike a track or in what contexts and why people might skip a track when listening to music using Shuffled Play.

(Implications for future research: Page 31, lines 651-655)

Reviewer #3

I agree that this is a novel method that provides some new insights -- in particular, insights regarding the importance of associations of tracks to information outside of the music itself. See however comments below regarding a potential distortion with regard to interpretation of this finding

Regarding the reporting of data -- I found it curious that the authors did not take advantage of the opportunity to describe differences between musicians and non-musicians -- this data appears to be available

We did investigate this but found no differences between musicians and non-musicians. We added a filter to the visualization so readers could explore this for themselves.

Differences between people who use music in their professional life and those that do not were investigated. There were no differences found in people responses to the closed questions and the themes identified from the open-ended questions.

(RQ2: Page 18, lines 347-350)

Lines 235-236. Using Tableau, I wasn't sure whether the modifications I was making to the filters were having an impact on the visualized data because of the proportion of each subpopulation in the total sample or whether it reflected the unique views of the population remaining after filtering. I wonder if using proportions would be more informative.

We are unsure to which particular visualisation you are referring, but in all cases, the filtering is configured by default (though it can be changed with a right click on it) to remove the individuals/responses that are not ticked. We agree that by using proportions it would be easier to see how relative values change between classes when some individuals are removed. We have made this change in the labels, while still maintaining the counts in the axis.

Lines 254-255, Possible missing details -- change to? "some of the participants who responded NO then went on to describe ..."

Thank you this has now been changed.

Interestingly, some of those participants who answered “no” describe aspects that fit with our understanding of the term ‘relationship’ within their responses (e.g. enjoyment, used for a specific purpose, associations etc.).

(Survey Experience: Page 16, lines 288-289)

Lines 313-315. for me at least, "relationship with a track" does not connote anything about emotion, regardless of what the piece is -- this is about my history with the band, song, or my associations with it -- hence, I don't see this contradicting the quantitative ranked data regarding "why I choose to listen". This interpretation seems like a bit of a distortion. I believe this point appears in different places in the manuscript --- abstract and discussion -- should be reconciled throughout.

Thank you for this comment. We have tried to make our point clearer within the manuscript. When preparing the survey we had the same assumption as you. However, when looking at the data, we found that most participants did not share this same assumption. For example, if “relationship with this track” is not about emotion and rather about associations then most people who see this questions would answer Yes and then describe how they always liked the band or remember hearing it with their friends. However this was not the case. About half the people answered with responses like: “No I don’t but I have always liked the band and I remember hearing it with my friends.”

For the question, “Do you have a relationship with this track? Whether it's 'Yes' or 'No' please explain”, participants responded that they did not have a relationship with a large proportion (about 50%) of the 784 total tracks discussed. Interestingly, some of those participants who answered “no” describe aspects that fit with our understanding of the term ‘relationship’ within their responses (e.g. enjoyment, used for a specific purpose, associations etc.). With some of the participants responding “no”, our findings suggest that our understanding of the word ‘relationship’ in the context of music listening may not be representative of how a substantial group of our respondents think about the music they listen to, associating this phrase with an emotional response rather than an associational one.

(Survey Experience: Page 16, lines 286-294)

Lines 455-458. the authors note that the ability to get insights about music that listeners do not like is a unique opportunity afforded by the shuffle methodology. but this is hardly the most effective means of getting this info - - they could have asked people to name and defend tracks they'll especially dislike. The authors should consider emphasizing that this is more of a "fringe" benefit of the method rather than something it is ideally designed to address

We agree there are better ways to understand people’s responses to music they do not like and agree this is a fringe benefit of this method. We have been more explicit that the method allows for a variety in responses which may include some negative ones.

Finally, by using the Shuffled Play method, we were able to elicit a variety in people’s responses. This included some tracks participants did not enjoy or did not like specific aspects of. This helps highlight the nuanced way in which people talk about music.

(Results Summary: Page 25, lines 503-506)

Reviewer #4:

"Do the Shuffle: Using shuffled play to explore reasons for music listening" is an exciting title that piqued my curiosity. A laudable contribution to research in music on why people download and listen to music, and which aspects of music listening experiences are people prioritize while listening to music on their devices.

The abstract is clear and answers the purpose of the study. I am curious about the tact that people download music because they like it – Seems to highlight the emotion of liking. Earlier, the authors state that participants use "sound and musical features" to characterize music and associate the music with their context "over ... emotional responses". "Therefore, how their prioritize music seems to contradict why the music is downloaded in the first place — an interesting dilemma.

Thank you for this interesting observation. We interpret your point here as being based on the assumption that liking music is an emotional response in itself. In this paper we do not start from this assumption, rather liking is taken to be an aesthetic judgment rather than an emotional response. See also, for example, Juslin’s framework (2013): “the present framework postulates that an aesthetic judgment can lead to both liking (i.e., preference) and emotion. Liking (or disliking) is a mandatory outcome, whereas emotion is a possible additional outcome.”

The introduction: A well-posited argument, the authors emphasize the importance of self-chosen music for research, music fit and context, functions of music listening, and music preference. They address the gaps in previous studies – music is preselected for participants — the importance of self-chosen music in the new technological era. Arguing that listeners today "carry their music with them," the authors use Shuffle Play an app that allows for the flexibility of listening choices, to collect data. The App is distinct because it places ownership in the hands of the listener while also allowing for a chance. In their introduction, the authors defend their arguments answering why they chose to research self-directed listening in the first place.

Aims and Questions: Aim to use a "mixed methods approach" to answering two questions: first, why people download and listen to music, and second, which aspects of the music listening experiences are prioritized by people listening to music on their devices.

Methods: Design: Even though the researchers tell us that they are using a mixed-methods approach, we do not know what kind of mixed methods approach. In mixed-methods, researchers collect both quantitative and qualitative data. Therefore, the plan may have combined methods, but the design is either, for example, Explanatory, sequential, convergent, or complex with embedded core designs. I encourage the researchers to read: Creswell, J. W., & Creswell, J. D. (2018). Research design: Qualitative, quantitative, and mixed methods approaches (5th ed.). Thousand Oaks, CA: Sage. They are using only a survey questionnaire for this purely descriptive study. Are the authors suggesting that Shuffle play is the qualitative part of their data collection?

Thank you for this helpful reference. You comment, along with that of reviewer 1 has helped us better describe our method. We have adjusted the manuscript to talk about the study as using mixed response type data.

With this method, along with collecting mixed response type data, we build on and go beyond the more typical foci in the literature on music and emotion, music in everyday life and music and technology.

(Aims and research questions: Page 7, lines 140-141)

For this descriptive study, we collect responses from a mix of question response types, including closed and open questions.

(Design: Page 7, lines 152-153)

Participants: The authors have a robust sample size (n.322). Excellent. It will be interesting to find out why 2/3rds are female, not sure how the authors will discuss this later.

We were aware of this during recruitment and we tried to work to recruit more men, through snowball sampling. However, we found this difficult. We now discuss this as a possible limitation in the limitation section. While this is a limitation it is not uncommon to have more women respondents for surveys (e.g. Curtin, R., Presser, S., & Singer, E. (2000). The effects of response rate changes on the index of consumer sentiment. Public Opinion Quarterly 64: 413–428.)

In addition, the participants themselves, which primarily represents female, young adults, and people from or residing in the UK, cannot be assumed to generalise to all contexts and cultures.

(Limitations Page 30, lines 621-623)

Interesting that they came from different countries. How was the sample accessed? Were they invited to participate? From a conference. Not sure about the representative sample argument – if we have single-digit representations from some countries. Using snowballing techniques from authors networks (See Procedure) does not make this representative.

We have tried to better explain that the initial networks invited to participate were from the authors’ personal and professional networks and were recruited via email and social media. We cannot find where we called the sample representative as this is a convenience sample. Please see our edits to the procedure section as we hope this now is clearer.

A snowball sampling (30,31) recruitment procedure was used with direct email invitations sent to target participants from the authors' professional and social networks.

(Procedure, Pages 13, lines 227-228)

Shuffle play: This has loads of potential for future research. I am not sure if I can call this qualitative data collection, primarily as a questionnaire, is also used to ask descriptive questions.

We agree and now have changed the methods section to better represent what was done.

With this method, along with collecting mixed response type data, we build on and go beyond the more typical foci in the literature on music and emotion, music in everyday life and music and technology.

(Aims and research questions: Page 7, lines 140-141)

For this descriptive study, we collect responses from a mix of question response types, including closed and open questions.

(Design: Page 7, lines 152-153)

Results: Associations, as the most influential type of response then, describing characteristics of a soundtrack and finally evaluation of the participant relationship, or familiarity with the music. How participants responded to music-related directly to their emotions – that is, they liked it.

My final thoughts: While this is a very promising study, it might be best for the authors to position it as a descriptive study using two forms of data collections- one through direct survey and other through online survey within a listening app. More work is needed to support the explanation of your methods.

Thanks again for pointing this out. This has hopefully been addressed in our response to the previous comments.

Attachment

Submitted filename: Response to Reviewers_FINAL.docx

Decision Letter 1

Sarah EP Munce

16 Jan 2020

Do the Shuffle: Exploring reasons for music listening though shuffled play

PONE-D-19-17641R1

Dear Dr. Spiro,

We are pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it complies with all outstanding technical requirements.

Within one week, you will receive an e-mail containing information on the amendments required prior to publication. When all required modifications have been addressed, you will receive a formal acceptance letter and your manuscript will proceed to our production department and be scheduled for publication.

Shortly after the formal acceptance letter is sent, an invoice for payment will follow. To ensure an efficient production and billing process, please log into Editorial Manager at https://www.editorialmanager.com/pone/, click the "Update My Information" link at the top of the page, and update your user information. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to enable them to help maximize its impact. If they will be preparing press materials for this manuscript, you must inform our press team as soon as possible and no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

With kind regards,

Sarah E.P. Munce, PhD

Academic Editor

PLOS ONE

Additional Editor Comments (optional):

Reviewers' comments:

Acceptance letter

Sarah EP Munce

31 Jan 2020

PONE-D-19-17641R1

Do the Shuffle: Exploring reasons for music listening through shuffled play.

Dear Dr. Spiro:

I am pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department.

If your institution or institutions have a press office, please notify them about your upcoming paper at this point, to enable them to help maximize its impact. If they will be preparing press materials for this manuscript, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org.

For any other questions or concerns, please email plosone@plos.org.

Thank you for submitting your work to PLOS ONE.

With kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Sarah E.P. Munce

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 File. Full questionnaire.

    Administered online through Qualtrics.

    (PDF)

    S2 File. Code book.

    Finalized code book used by both coders.

    (PDF)

    S3 File. Track and extracted Spotify data.

    Track information given by the participant and track data extracted from “Sort Your Music” application.

    (PDF)

    Attachment

    Submitted filename: Response to Reviewers_FINAL.docx

    Data Availability Statement

    All of the raw data is freely available and can be downloaded for reuse from [https://miguems.github.io/dotheshuffle/], a webpage specifically designed to accompany this paper.


    Articles from PLoS ONE are provided here courtesy of PLOS

    RESOURCES