Skip to main content
PLOS One logoLink to PLOS One
. 2025 Sep 11;20(9):e0329711. doi: 10.1371/journal.pone.0329711

EEG Sonification improves sleep staging performance in novice stagers

Sam Chin 1,#, Nathan W Whitmore 2,*,#, Nathan Perry 1, Joe Paradiso 1, Pattie Maes 2
Editor: Hesam Ramezanzade3
PMCID: PMC12425293  PMID: 40934291

Abstract

Sleep staging is a critical tool used in research and clinical settings to evaluate and diagnose sleep conditions; however, sleep staging is labor intensive and may be challenging for inexperienced practitioners. We explored whether adding an auditory representation (sonification) of the EEG to a standard visual representation could improve sleep staging performance or reduce workload. This is the first study to investigate the effects of sonification on sleep staging performance. We performed a within-subjects study in which 40 participants completed an online sleep staging task with and without sonified EEG. EEG was sonified by minimal transformation in which the raw EEG signal was played as an audio signal. Contrary to our hypothesis, we found adding sonification did not result in improvements in accuracy, speed, or workload for the entire subject group. However, when we stratified participants by sleep staging experience, we found sonification improved accuracy for the least experienced participants. These findings suggest EEG sonification may be useful as a tool to enable novice sleep stagers to reach acceptable performance levels faster.

1 Introduction

During a night of sleep, humans transition through distinct physiological states, referred to as sleep stages. Sleep staging is a medical diagnostic procedure in which a technician classifies sleep into distinct sleep stages using a standardized set of rules applied to EEG, electromyogram, and electroculogram data [1].

Our study aimed to examine whether adding an audio representation of the EEG to a sleep staging interface could improve sleep staging performance. Prior research has investigated developing algorithms to convert EEG into sound [25], but has not yet addressed whether this technique can improve accuracy or reduce workload of sleep staging.

1.1 Manual sleep staging is challenging

Sleep staging is a frequently performed medical diagnostic procedure. In the United States, it is performed more than a million times per year in order to test for conditions such as sleep apnea [6]. These stages are then used to compute clinical sleep indices such as sleep onset time, wake after sleep onset, and sleep fragmentation to diagnose specific sleep conditions such as apnea and periodic limb movement disorder [1,7].

Manual sleep staging is performed by examining electrical recordings of brain activity (EEG), eye movements (electrooculogram, EOG), and facial muscle tone (EMG) during the sleep period in 30-second epochs. Sleep scorers apply a standardized algorithm to classify each segment as Wake, REM sleep, or non-REM sleep stages 1-3. Sleep stages are identified using discrete visual visual features in these signals such as blinks, rapid eye movements, muscle twitches, and shifts in the frequency and waveform of the EEG. For example, wake is defined by the presence of 10-Hz oscillations in the EEG which are absent in sleep, REM sleep is defined by muscle paralysis and frequent rapid eye movements, and N1-N3 are associated with increasing amplitude of delta brain waves and the occurrence of specific waveforms like K-complexes and sleep spindles [1].

Manual sleep staging is highly labor intensive, and requires specialized training [8]. Furthermore, even highly trained sleep stagers show agreements only around 83% [8]. Thus, technology to improve sleep staging performance has the potential to improve both sleep staging accuracy and efficiency.

While machine learning algorithms have demonstrated good performance in sleep staging, [9] they are not used extensively in clinical practice due to concerns about their reliability and inability to provide justifications for their decisions. Regulatory agencies like the FDA have emphasized that medical AI algorithms should include a human-in-the-loop to verify decisions [10], thus manual sleep staging will likely continue to be necessary despite advances in AI sleep staging. It is therefore important to develop technologies which can facilitate faster and more accurate manual sleep staging.

1.2 EEG sonification for sleep staging

A potential approach for improving human judgment in sleep staging is EEG sonification, or presenting an auditory representation of the EEG. We take a multisensory approach and present the user with both an auditory and visual representation, since these can be integrated in the brain and improve perception of a stimulus [11,12].

There is significant prior art in sonifying bioelectric signals for analysis [13,14]. Some of the earliest work on neurons used an auditory interface to monitor activations [15]. Previous research has shown that sonification can reduce the time required to classify brain activity as epileptic or normal, suggesting that sonification can reduce the workload of EEG interpretation [2,3]. Research has also examined developing sleep-EEG sonification algorithms which balance aesthetic qualities with retaining important information in the EEG [4,5]. However, to date, no research has examined if adding sonification to an existing visual display system for sleep staging can improve accuracy or reduce workload of sleep staging.

1.3 How could sonification improve performance and reduce workload?

The brain combines information from sensory modalities (vision, audition, haptic, etc.) to form a unified percept of the world using multisensory integration. Multisensory integration maximizes reliability and minimizes variance of the integrated percept [16]. Having a multisensory stream of information allows each sense to function where it is most effective. For example, audition is more temporally precise, while vision is more spatially precise [17,18]. According to the Multiple Resource Model, a multisensory presentation may also reduce workload by moving parts of a complex visual task to an alternate sensory modality [19]. Sleep staging may be especially likely to benefit from multisensory representation because distinguishing sleep stages requires discriminating fine differences in the dominant frequency of signals, which may benefit from the temporal and frequency precision of audition. In general by presenting both an auditory and visual signal, the brain can select the sensory modality with the most information.

For example, in a loud room, we use lip reading to augment our auditory perception of speech [20]. By combining both visual and auditory information, the brain creates a more reliable percept than either modality alone. This is also confirmed by experimental results which have demonstrated that sound can enhance visual perception [11,12,21].

Humans naturally learn to use multisensory integration implicitly, when useful information occurs in two or more sensory modalities. For example, through repeated exposure to the sight of objects falling and the accompanying sound, we develop an intuitive understanding of the physics of falling objects [22]. Adding sonification to a task has been demonstrated to assist in learning perceptual and motor skills [23,24] and interpretation of EEG data [2,3]. Therefore, mere exposure to sonified EEG (along with feedback) could potentially enable participants to learn to use sonic features to improve their performance sleep staging performance.

1.4 Study design

We tested whether sonifying EEG could improve accuracy, speed, and cognitive load using an online sleep staging task where participants were asked to stage sleep either with or without sonification. Specifically, we aimed to test the following hypotheses:

  • H1: Participants will perform sleep staging more accurately (as quantified by Cohen’s kappa agreement with gold standard scorers) when they view an EEG with an accompanying sonified representation compared to viewing the EEG alone.

  • H2: Participants will stage faster when they view EEG and also hear a sonified representation than when they view the EEG without sonification.

  • H3: Participants will report lower mental demand, effort and frustration as measured by the NASA Task Load Index [25] when staging with sonification as compared to visual staging alone.

  • H4: Benefits will be selective for sleep epochs which contain alpha waves or sleep spindles.

2 Materials and methods

We preregistered our procedure prior to analysis at https://osf.io/2xz5e. Our analysis followed the procedures laid out in the preregistration with one exception. We chose not to analyze the number of transitions between epochs since fewer than 5 participants had transitions. The research was exempted from review by the MIT IRB (protocol #E-4933 ) and written consent was obtained online.

2.1 Participants and screening procedure

We recruited participants from March 8 2024 to March 18 2024 via advertisements distributed to sleep researchers on Twitter. Participants were eligible if they were at least 18 years of age, and reported some experience with AASM sleep staging. After observing a large number of probable bots attempting to complete the survey we also added an additional screening question requiring participants to correctly identify an obvious epoch of Wake to proceed; this question was added after the 386th participant. 536 participants completed screening and proceeded to the practice test.

To exclude inattentive participants and bots, we excluded participants with implausibly short reaction times in the practice block. Participants were not allowed to proceed to the test blocks if their median time to stage an epoch during the practice block was less than 2 s. Any participants who failed this criterion were excluded from the data. The criterion was selected based on the reaction time distribution of 11 manually identified likely bots (μ of medians 0.59 s, σ = 0.27 s) vs 3 known valid sleep stagers (μ of medians 5.5 s, σ = 2.1 s).

There were 623 attempts to complete the practice block; of these, 46 met the reaction time criterion and proceeded to the test blocks. 40 of these participants completed all of the test phases and were included in analysis. The included participants consisted of 24 males, 15 females, and 1 non-binary person with a mean age of 24 ± 3.11 years. Reported sleep staging experience is shown in Table 1.

Table 1. Count of participants by sleep staging experience level.

Number of Polysomnograms Staged in Past Year Count
1-10 14
11-20 8
21-40 10
41-80 5
More than 80 3

2.2 Stimulus generation procedure

We used data from healthy participants in the ISRUC-SLEEP database [26] downloaded on 27/02/2024. We included only sleep epochs which were staged as the same sleep stage by both ISRUC-SLEEP stagers. For more details on the sleep dataset, see [26].

We created the visual and auditory representations of the data using a modified version of Visbrain Sleep [27], an open-source sleep staging tool. An example of the visual display is shown in Fig 1. The data were displayed in an AASM standard format [1] for sleep staging with 5 channels displayed: 3 EEG channels (F3, C3, O1), 1 electroculogram channel (left outer canthus), and 1 chin EMG channel. All data were referenced to the right ear. All channels were scaled to ± 100 μV. Participants viewed one 30 second-long epoch of sleep at a time.

Fig 1. An example of the online sleep staging task interface used by participants.

Fig 1

The O1-A2 channel was sonified using our stimulus generation procedure.

To generate the audio data we performed a minimal transformation where we converted 30 seconds of visually displayed data from the O1 channel into a 1.36 second sound clip. We used this approach rather than a more complex transformation to avoid destroying any information and to maximize participants’ ability to learn. We chose a simple transformation over a musical or mapping-based approach because these risk increasing the salience of features irrelevant to sleep staging. A simple transformation allows the brain’s auditory system to naturally identify the relevant properties of the sound. This approach enables the formation of perceptual boundaries based on the sound’s intrinsic characteristics [28]. The original dataset was 30 s epochs measured at 200 Hz [29]. This was downsampled to 100 Hz in Visbrain. In order to produce the auditory stimulus, we sped up the visual data by approximately 95% for a final audio length of 1.36 s. We also performed scaling and padding of the audio to reduce clipping and popping during playback. We observed that this method produced qualitatively distinct sounds for each sleep stage. The resulting audio samples were low-frequency (20–1000 Hz) and were mixed to -12 dB (approximately standard loudness for musical audio tracks). Examples of the sounds (along with accompanying visual representation) can be found at https://github.com/nathanww/SoundspindleExperiment/tree/main/data/.

2.3 Experimental procedure

On beginning the experiment, participants first completed a questionnaire on their age, sex, and sleep staging experience. To measure experience, participants were asked how many polysomnograms they had scored in the past year with five possible choices: 1-10, 11-20, 21-40, 41-80, or more than 80 (Table 1). We choose to measure a “snapshot” of scoring in the past year (rather than lifetime scoring) as we predicted participants would be more accurate at estimating scoring over past year.

Participants then proceeded to the online sleep staging task summarized in Fig 2. The task began with an orientation to the display format (Fig 1), after which participants began the practice block of sleep staging. In this block, participants viewed and attempted to stage 20 30-second epochs of sleep. Participants were also able to toggle between the epoch to be staged, and the 30 seconds before and after this epoch as many times as desired before issuing a sleep stage. Sound corresponding to the O1 channel of the EEG was played immediately upon viewing the sleep epoch and any time the participant moved forward or backward in time.

Fig 2. The experimental procedure.

Fig 2

After the participant issued a sleep stage, the sonification of the O1 channel was played again and the participant received feedback on the correct sleep stage. The participant could then proceed to the next epoch of sleep. Epochs were presented in random order until the participant issued the correct response; the epoch was then dropped from the rotation. Once all epochs were correctly identified, the block ended.

Participants then proceeded to the first and second test blocks, which presented information identically to the practice block. Each included 30 epochs of sleep, and each epoch was shown only once, regardless of whether it was correctly staged. Participants were randomized and counterbalanced such that half received sonification on the first test block (but not the second) and half received sonification on the second test block (but not the first).

At the end of each test block, participants also performed the NASA Task Load Index [25] to assess cognitive load while performing the task. Following completion of the last block, we asked participants for their overall impressions on the experiment.

2.4 Statistical analysis

For all statistical analyses, we compared performance in the no-sound block to performance in the sound block using a paired test. We used a Wilcoxon signed-rank test for all comparisons except for the comparison of task load index scales, where the differences between blocks met normality assumptions for using a paired t-test. For all analyses, we defined the correct sleep stage for each epoch as the stage given by the ISRUC stagers.

We measured sleep staging accuracy using Cohen’s kappa to measure agreement between the participant’s stages and the correct stages in each block. Cohen’s kappa provides an agreement score from -1 (complete disagreement) to 1 (perfect agreement) corrected for the agreement expected from random chance [30]. Importantly, kappa compensates for the highly imbalanced classes found in sleep staging which can distort simpler measurements of percentage correct.

To determine whether participants’ sleep staging experience mediated the effect of sound, we split participants into 5 categories based on the number of polysomnograms they reported staging in the prior year. We then performed a paired test comparing kappa for the sound vs no-sound block for each experience group.

When analyzing reaction times, we only included trials where the participant gave the correct sleep stage. We excluded trials with outlier reaction times (RTs) more than two sigma above or below the average reaction time consistent with standard practices [31].

We hypothesized that sonification might selectively improve the ability to identify Wake and N2, which were associated with especially distinctive sounds. To determine if sonification improved participants’ ability to identify specific sleep stages, we defined the true stage for each epoch as the stage given by the ISRUC stagers. We then measured the percent of “true stage X” that was identified as that stage by participants. For example, if participants recognized 5 of the 10 Wake epochs present in a block, their accuracy for Wake would be 50%. For this analysis, we quantified participant performance using percent correct rather than Cohen’s kappa because kappa is undefined when the correct sleep stage is the same for all epochs.

We also computed how sonification affected the time required to identify stage Wake and stage N2 specifically (Fig 2). For these analyses, we only included trials where the RT was within 2 sigma of the mean and the sleep stage was correctly identified; 3 participants failed to correctly identify any instances of stage Wake in one or more blocks and were excluded from this analysis.

3 Results

3.1 Sonification improves staging accuracy for participants with least experience

We found no significant difference between sleep staging performance in the sound and no-sound block [Wilcoxon signed-rank test, p = 0.21], and therefore no support for H1. Similarly we did not find support for H2, as the time to complete sleep staging did not differ significantly between the sound and no-sound blocks for all epochs staged [Wilcoxon signed-rank test, p = 0.65] or epochs staged correctly [Wilcoxon signed-rank test, p = 0.83].

However, our exploratory analysis revealed that sound improved accuracy measured by kappa only for the group of participants who reported having staged 1–10 nights of sleep previously; for this subgroup kappa was significantly higher with sound [Wilcoxon signed-rank test, p = 0.01] (Fig 3). No other group of participants showed a significant effect of sound.

Fig 3. (Left) Sleep staging performance (Cohen’s kappa) of participants in the sound and no-sound blocks by experience level.

Fig 3

(Right) Difference in sleep staging performance between the sound and no-sound blocks. The least experienced participants showed a significant improvement in kappa with sound (indicated by *). Higher kappa indicates better staging performance.

3.2 Sonification did not significantly alter task load

To test our prediction that sonification would reduce the demand of sleep staging [H3], we measured whether NASA Task Load Index of mental load, effort, and frustration differed between the sound block and the no-sound block. As the differences were normally distributed, we used a paired t test. We did not observe any significant differences between the sound and no-sound conditions on any of the three axes (Fig 4). We also did not observe significant effects on the TLX for the subgroup of least experienced users (who reported scoring 1-10 nights of sleep).

Fig 4. The effect of sonification on NASA TLX scores was not statistically significant.

Fig 4

Mental Demand [t(39) = 1.052, p = 0.30], Physical Demand [t(39) = 0.524, p = 0.60], Temporal Demand [t(39) = 1.622, p = 0.11], Performance [t(39) = 0.826, p = 0.41], Effort [t(39) = 0.434, p = 0.67], Frustration [t(39) = -0.510, p = 0.61].

3.3 Effects of sonification did not depend on sleep stage

We predicted [H4] that sonification might selectively improve the recognition of Wake and N2 as those sleep stages contain features that are particularly identifiable in sound. However, we found that neither accuracy in identifying Wake nor accuracy in identifying N2 was improved by sonification (Fig 5), [Wilcoxon signed-rank test | Wake: p = 0.57, N2: p = 0.28]. Therefore we did not find support for this hypothesis.

Fig 5. Effect of sonification on accuracy for each sleep stage.

Fig 5

The correct sleep stage for an epoch was defined by the ISRUC sleep stagers. Accuracy by sleep stage was calculated as the number of epochs in the sleep stage that were correctly identified by the participant. We did not observe significant differences in the effect of sonification by sleep stage.

3.4 Sleep staging experience predicts test performance

We observed a significant correlation between the number of polysomnograms staged in the past year and the overall kappa across both test blocks [r(38) = 0.49, p = 0.001].

3.5 Participants found sonification useful

After completing the experiment, 30/40 participants reported some benefit from the sound in staging sleep. 26 participants reported that the sound was “Useful, a lot” for staging, while 4 participants reported it was “Useful, a little”, 4 indicated that they were unsure, and 6 reported that it was not useful. 39/40 participants indicated that they would use a sleep staging platform that incorporated sonification.

4 Discussion

In this study, we found that adding sonification to an EEG sleep staging task improved sleep staging accuracy only for the group of participants with the least experience. Contrary to predictions, we did not find that sonification reduced the time to complete staging or the mental load of the staging task; though participants reported finding the sounds useful.

Why did we find sonification improved performance in novices, but not in experienced sleep stagers? One hypothesis is that there may be a ceiling effect where experienced sleep stagers can extract all the information in the signal from its visual representation; thus sound does not add any additional information. An alternative hypothesis is that experienced sleep stagers may experience a blocking effect where well-learned associations between visual features and sleep stages prevent effective learning of the relationship between sound features and sleep stage. Blocking is a well-documented effect in learning where once people reach good performance on a task they fail to incorporate other useful information because of a lack of prediction errors [32,33]).

Importantly, prior work on the effects of sonification on task performance [2,3,21,24] have typically used tasks that are at least somewhat novel to participants, where a blocking effect would not typically be observed. In contrast, the majority of our participants had extensive experience with visual sleep staging. This difference in prior experience (along with blocking or ceiling effects) why we did not observe the large sonification benefits observed in prior studies.

While we did not find overall significant effects of sonification, participants did frequently report that sound was helpful for sleep staging, and we observed non-significant improvements in the time to complete sleep staging. It is possible that more experienced participants would also show benefits from sonification with a larger sample size or with a longer period of learning.

Future studies could further explore the benefits of sonification for novice sleep stagers by comparing the rates at which naive participants improve in sleep staging when trained with visual plus sonified information compared to participants training with visual information alone. Because blocking effects should not occur when there is no pre-existing knowledge, novice participants trained with sonification may learn faster and attain higher levels of performance than those trained with visual information alone. Future studies could also explore the effects of alternative sound transformations or whether an increased number of learning trials improves the effects of sonification.

As a compensated online study, there is a risk of participants engaging in deception to collect rewards; risks here include both automated and semi-automated bots and human participants who do not perform the task according to instructions [34]. We employed multiple measures to detect and exclude these response types, including reCAPTCHA, excluding participants who failed to identify an obvious epoch of Wake, and excluding participants who exhibited implausibly fast responses during the practice block. We also demonstrated that the reported level of sleep staging experience correlated with performance, suggesting that the bulk of our participants correctly reported their experience level and attempted to perform the sleep staging task. Nonetheless, it is possible that the recorded answers contain some non-genuine responses.

A unique feature of this study compared to prior work on sonification of bioelectric signals is that we examined whether sonification combined with standard techniques could augment performance. In contrast, previous studies have examined the optimal techniques to create aesthetically pleasing sounds [4,5] or whether sonification could replace visual interpretation [2,3]. We found that sonification improved accuracy in novice sleep stagers and was generally found to be helpful by most users. Our observation that sonification can increase sleep staging accuracy in inexperienced sleep stagers suggests that it may be particularly valuable for non-specialists with limited sleep experience (such as internists or students). Future research should also explore whether augmenting sleep staging training programs with sonification can speed learning and improve performance.

Data Availability

Data and code used to perform all the analyses in this study are available at https://doi.org/10.5061/dryad.3bk3j9kz0.

Funding Statement

The author(s) received no specific funding for this work.

References

  • 1.Berry RB, Brooks R, Gamaldo CE, Harding SM, Marcus C, Vaughn BV. The AASM manual for the scoring of sleep and associated events. rules, terminology and technical specifications. Darien, Illinois: American Academy of Sleep Medicine; 2012.
  • 2.Gomez-Quintana S, O’Shea A, Factor A, Popovici E, Temko A. A method for AI-assisted human interpretation of biological signals: analysis of neonatal EEG; 2022. [DOI] [PMC free article] [PubMed]
  • 3.Hobbs K, Krishnamohan P, Legault C, Goodman S, Parvizi J, Gururangan K, et al. Rapid bedside evaluation of seizures in the ICU by listening to the sound of brainwaves: a prospective observational clinical trial of Ceribell’s brain stethoscope function. Neurocrit Care. 2018;29(2):302–12. doi: 10.1007/s12028-018-0543-7 [DOI] [PubMed] [Google Scholar]
  • 4.Moradi F, Mohammadi H, Rezaei M, Sariaslani P, Razazian N, Khazaie H, et al. A novel method for sleep-stage classification based on sonification of sleep electroencephalogram signals using wavelet transform and recurrent neural network. Eur Neurol. 2020;83(5):468–86. doi: 10.1159/000511306 [DOI] [PubMed] [Google Scholar]
  • 5.Franco P, Värri A. Experiments of the sonification of the sleep electroencephalogram. Finnish Journal of eHealth and eWelfare. 2015;7(2–3):65–74. [Google Scholar]
  • 6.Tachibana N, Ayas NT, White DP. A quantitative assessment of sleep laboratory activity in the United States. Journal of Clinical Sleep Medicine. 2005;01(01):23–6. doi: 10.5664/jcsm.26292 [DOI] [PubMed] [Google Scholar]
  • 7.Shrivastava D, Jung S, Saadat M, Sirohi R, Crewson K. How to interpret the results of a sleep study. J Community Hosp Intern Med Perspect. 2014;4(5):24983. doi: 10.3402/jchimp.v4.24983 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Rosenberg RS, Van Hout S. The American Academy of sleep medicine inter-scorer reliability program: sleep stage scoring. J Clin Sleep Med. 2013;9(1):81–7. doi: 10.5664/jcsm.2350 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Bandyopadhyay A, Goldstein C. Clinical applications of artificial intelligence in sleep medicine: a sleep clinician’s perspective. Sleep Breath. 2023;27(1):39–55. doi: 10.1007/s11325-022-02592-4 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Food US, Administration D, et al. Good machine learning practice for medical device development. 2021. https://www.fda.gov
  • 11.Vroomen J, de Gelder B. Sound enhances visual perception: cross-modal effects of auditory organization on vision. J Exp Psychol Hum Percept Perform. 2000;26(5):1583–90. doi: 10.1037//0096-1523.26.5.1583 [DOI] [PubMed] [Google Scholar]
  • 12.Kohlrausch A, van de Par S. Audio—visual interaction in the context of multi-media applications. Communication acoustics. Berlin, Heidelberg: Springer; 2005. p. 109–38. 10.1007/3-540-27437-5_5 [DOI]
  • 13.Cherston J, Paradiso JA. Rotator: flexible distribution of data across sensory channels. In: Proceedings of the 23rd International Conference on Auditory Display - ICAD 2017 . 2017. p. 86–93. 10.21785/icad2017.009 [DOI]
  • 14.Ballora M, Pennycook B, Ivanov PC, Glass L, Goldberger AL. Heart rate sonification: a new approach to medical diagnosis. Leonardo. 2004;37(1):41–6. doi: 10.1162/002409404772828094 [DOI] [Google Scholar]
  • 15.Hubel DH. Tungsten microelectrode for recording from single units. Science. 1957;125(3247):549–50. doi: 10.1126/science.125.3247.549 [DOI] [PubMed] [Google Scholar]
  • 16.Ernst MO, Banks MS. Humans integrate visual and haptic information in a statistically optimal fashion. Nature. 2002;415(6870):429–33. doi: 10.1038/415429a [DOI] [PubMed] [Google Scholar]
  • 17.Witten IB, Knudsen EI. Why seeing is believing: merging auditory and visual worlds. Neuron. 2005;48(3):489–96. doi: 10.1016/j.neuron.2005.10.020 [DOI] [PubMed] [Google Scholar]
  • 18.Ramezanzade H, Abdoli B, Farsi A, Sanjari MA. The effect of sonification of visual modeling on relative timing: the role in motor learning. Acta Kinesiologica. 2017;11:17–27. [Google Scholar]
  • 19.Wickens CD. Multiple resources and mental workload. Hum Factors. 2008;50(3):449–55. doi: 10.1518/001872008X288394 [DOI] [PubMed] [Google Scholar]
  • 20.Goldstein EB, Cacciamani L. Sensation and perception. 11th ed. Boston, MA: Cengage; 2022.
  • 21.Effenberg AO. Movement sonification: effects on perception and action. IEEE Multimedia. 2005;12(2):53–9. doi: 10.1109/mmul.2005.31 [DOI] [Google Scholar]
  • 22.Battaglia PW, Hamrick JB, Tenenbaum JB. Simulation as an engine of physical scene understanding. Proc Natl Acad Sci U S A. 2013;110(45):18327–32. doi: 10.1073/pnas.1306572110 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23.Effenberg AO, Fehse U, Schmitz G, Krueger B, Mechling H. Movement sonification: effects on motor learning beyond rhythmic adjustments. Front Neurosci. 2016;10:219. doi: 10.3389/fnins.2016.00219 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24.Parimi F, Abdoli B, Ramezanzade H, Aghdaei M. The effect of internal and external imagery on learning badminton long serve skill: the role of visual and audiovisual imagery. PLoS One. 2024;19(9):e0309473. doi: 10.1371/journal.pone.0309473 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Hart SG, Staveland LE. Development of NASA-TLX (task load index): results of empirical and theoretical research. In: Advances in Psychology. vol. 52. Elsevier; 1988. p. 139–83. https://linkinghub.elsevier.com/retrieve/pii/S0166411508623869 [Google Scholar]
  • 26.Khalighi S, Sousa T, Santos JM, Nunes U. ISRUC-sleep: a comprehensive public dataset for sleep researchers. Comput Methods Programs Biomed. 2016;124:180–92. doi: 10.1016/j.cmpb.2015.10.013 [DOI] [PubMed] [Google Scholar]
  • 27.Combrisson E, Vallat R, O’Reilly C, Jas M, Pascarella A, Saive A-L, et al. Visbrain: a multi-purpose GPU-accelerated open-source suite for multimodal brain data visualization. Front Neuroinform. 2019;13:14. doi: 10.3389/fninf.2019.00014 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.Ananthabhotla I. Cognitive audio: enabling auditory interfaces with an understanding of how we hear. Massachusetts Institute of Technology; 2022. https://dspace.mit.edu/handle/1721.1/143241
  • 29.Khalighi S, Sousa T, Santos JM, Nunes U. ISRUC-sleep: a comprehensive public dataset for sleep researchers. Comput Methods Programs Biomed. 2016;124:180–92. doi: 10.1016/j.cmpb.2015.10.013 [DOI] [PubMed] [Google Scholar]
  • 30.McHugh ML. Interrater reliability: the kappa statistic. Biochem Med (Zagreb). 2012;22(3):276–82. doi: 10.11613/bm.2012.031 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 31.Berger A, Kiefer M. Comparison of different response time outlier exclusion methods: a simulation study. Front Psychol. 2021;12:675558. doi: 10.3389/fpsyg.2021.675558 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.Kamin LJ. Attention-like processes in classical conditioning. TR-5; 1967.
  • 33.Luque D, Vadillo MA, Gutiérrez-Cobo MJ, Le Pelley ME. The blocking effect in associative learning involves learned biases in rapid attentional capture. Q J Exp Psychol (Hove). 2018;71(2):522–44. doi: 10.1080/17470218.2016.1262435 [DOI] [PubMed] [Google Scholar]
  • 34.Loebenberg G, Oldham M, Brown J, Dinu L, Michie S, Field M, et al. Bot or not? detecting and managing participant deception when conducting digital research remotely: case study of a randomized controlled trial. J Med Internet Res. 2023;25:e46523. doi: 10.2196/46523 [DOI] [PMC free article] [PubMed] [Google Scholar]

Decision Letter 0

Hesam Ramezanzade

3 Jun 2025

PONE-D-25-11016EEG Sonification Improves Sleep Staging Performance in Novice StagersPLOS ONE

Dear Dr. Whitmore,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please submit your revised manuscript by Jul 18 2025 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Hesam Ramezanzade, Ph.D

Academic Editor

PLOS ONE

Journal Requirements:

1. When submitting your revision, we need you to address these additional requirements. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf 2. Please update your submission to use the PLOS LaTeX template. The template and more information on our requirements for LaTeX submissions can be found at http://journals.plos.org/plosone/s/latex. 3. Thank you for uploading your study's underlying data set. Unfortunately, the repository you have noted in your Data Availability statement does not qualify as an acceptable data repository according to PLOS's standards. At this time, please upload the minimal data set necessary to replicate your study's findings to a stable, public repository (such as figshare or Dryad) and provide us with the relevant URLs, DOIs, or accession numbers that may be used to access these data. For a list of recommended repositories and additional information on PLOS standards for data deposition, please see https://journals.plos.org/plosone/s/recommended-repositories.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Partly

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: Yes

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: No

Reviewer #2: Yes

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: This study is an interesting study on the effect of auditory intervention on sleep staging that has not been investigated so far. Most studies in the field of sanitization have used this type of intervention to develop (cognitive-motor) performance and have shown positive results.

1- The abstract is well-written; however, it would be beneficial to include a brief one-line reference to the feature that was presented to participants as an auditory pattern (sonification).

2- The introduction is well-structured. However, it would be beneficial to examine the effects of sonification on other functions, including motor performance, before discussing research related to sleep staging performance. Numerous studies have been conducted in this regard, some of which I have referenced. For esample:

Ramezanzade, H. (2019). Adding Acoustical to visual movement patterns to retest whether imitation is goal- or pattern- directed. Perceptual and Motor Skills, 127(1), https://doi.org/10.1177/0031512519870418.

Parimi F, Abdoli B, Ramezanzade H, Aghdaei M (2024) The Effect of Internal and external imagery on learning badminton long serve skill: The role of visual and audiovisual imagery. PLoS ONE 19(9): e0309473. https://doi.org/10.1371/journal.pone.0309473.

Effenberg AO, Fehse U, Schmitz G, Krueger B, Mechling H. Movement sonification: effects on motor learning beyond rhythmic adjustments. Frontiers in neuroscience. 2016;10:219. pmid:27303255.

3- You can reference the following source for the statement "audition is more temporally precise, while vision is more spatially precise".

Ramezanzade, H., Abdoli, B., Farsi, A., & Sanjari, M. A. (2017). The effect of sonification of visual modeling on relative timing: The role in motor learning. Acta Kinsiologica, 11, Supp. 2, 17-27.

4- You can also reference the following source for the statement “This is also confirmed by experimental results which have demonstrated that sound can enhance visual perception”.

Ramezanzade, H., Abdoli, B., Farsi, A., & Sanjari, M. A. (2017). The effect of sonification modelling on perception and accuracy of performing jump shot basketball. International Journal of Sports Studies, 4(11), 1388-1392.

Effenberg AO. Movement sonification: Effects on perception and action. IEEE multimedia. 2005;12(2):53–9.

5- Please specify precisely why the O1, F1, and C3 regions were selected for generating the auditory pattern.

6- The methodology is well-written, but it should include more details regarding the physical characteristics of the sound (frequency, volume, pitch, timbre, etc.)

7- It would be better to describe sleep staging precisely and in detail in both the introduction and methodology sections, while also specifying the characteristics of each stage.

8- The results section is well-reported. The discussion is also good, but it would be better to provide more speculation regarding the reasons behind the obtained results. It would be beneficial to study recent research related to sleep and its influencing variables, and to explain the results with the help of proposed hypotheses. The discussion is currently incomplete and inadequate.

Reviewer #2: Re: EEG Sonification Improves Sleep Staging Performance in Novice Stagers

General comments

Manuscript “EEG Sonification Improves Sleep Staging Performance in Novice Stagers” aimed at investigating whether adding an auditory representation (sonification) of the EEG to a standard visual representation could improve sleep staging performance. Additionally, the study examines if this method reduces workload. The authors found that adding sonification did not result in improvements in accuracy, speed, or workload for the entire subject group .However we found sonification improved accuracy for the least experienced participants. This study explores an interesting and promising concept. Below are some suggestions that should be considered by the authors of the manuscript:

Specific comments

Abstract

1. In the abstract, also mention the classification of groups based on the level of experience.

Methods

2. How many groups did you have in total? Please provide a detailed explanation of how the groups were divided based on the level of experience and the reception of the Sanifcation sound.

Introduction

1. The first paragraph of the “Manual Sleep Staging” section needs more references.

2. Hypothesis 1 should be written in full. In which aspect does it perform better?

Statistical analysis

1. Why is only the prior year experience of the participants considered? It seems that if the overall experience of the participants is taken into account, the accuracy of the study would be higher.

Discussion

Overall the discussion is not well written.

1. The consequences of the findings in relation to the hypotheses proposed in this study should be clarified with the existing literature.

2. How do you justify the effectiveness of Sanification for novices? You need to reference relevant articles on this.

3. Paragraph 5 is not relevant to this section. It is suggested that it be briefly mentioned in the Method section.

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

Attachment

Submitted filename: Re.docx

pone.0329711.s001.docx (13KB, docx)
PLoS One. 2025 Sep 11;20(9):e0329711. doi: 10.1371/journal.pone.0329711.r002

Author response to Decision Letter 1


19 Jul 2025

Thank you for your comments! We have improved the paper accordingly as described below.

Journal Requirements:

1. When submitting your revision, we need you to address these additional requirements.

Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. Please update your submission to use the PLOS LaTeX template. The template and more information on our requirements for LaTeX submissions can be found at http://journals.plos.org/plosone/s/latex.

We have updated the paper to use the PLoS Latex template and formatting style

3. Thank you for uploading your study's underlying data set. Unfortunately, the repository you have noted in your Data Availability statement does not qualify as an acceptable data repository according to PLOS's standards.

We have uploaded the data to Dryad

Reviewer 1 comments

Reviewer #1: This study is an interesting study on the effect of auditory intervention on sleep staging that has not been investigated so far. Most studies in the field of sanitization have used this type of intervention to develop (cognitive-motor) performance and have shown positive results.

1- The abstract is well-written; however, it would be beneficial to include a brief one-line reference to the feature that was presented to participants as an auditory pattern (sonification).

We have added this to the abstract

2- The introduction is well-structured. However, it would be beneficial to examine the effects of sonification on other functions, including motor performance, before discussing research related to sleep staging performance. Numerous studies have been conducted in this regard, some of which I have referenced. For esample:

Ramezanzade, H. (2019). Adding Acoustical to visual movement patterns to retest whether imitation is goal- or pattern- directed. Perceptual and Motor Skills, 127(1), https://doi.org/10.1177/0031512519870418.

Parimi F, Abdoli B, Ramezanzade H, Aghdaei M (2024) The Effect of Internal and external imagery on learning badminton long serve skill: The role of visual and audiovisual imagery. PLoS ONE 19(9): e0309473. https://doi.org/10.1371/journal.pone.0309473.

Effenberg AO, Fehse U, Schmitz G, Krueger B, Mechling H. Movement sonification: effects on motor learning beyond rhythmic adjustments. Frontiers in neuroscience. 2016;10:219. pmid:27303255.

3- You can reference the following source for the statement "audition is more temporally precise, while vision is more spatially precise".

Ramezanzade, H., Abdoli, B., Farsi, A., & Sanjari, M. A. (2017). The effect of sonification of visual modeling on relative timing: The role in motor learning. Acta Kinsiologica, 11, Supp. 2, 17-27.

4- You can also reference the following source for the statement “This is also confirmed by experimental results which have demonstrated that sound can enhance visual perception”.

Ramezanzade, H., Abdoli, B., Farsi, A., & Sanjari, M. A. (2017). The effect of sonification modelling on perception and accuracy of performing jump shot basketball. International Journal of Sports Studies, 4(11), 1388-1392.

Effenberg AO. Movement sonification: Effects on perception and action. IEEE multimedia. 2005;12(2):53–9.

We have added the Parimi at al (2024), Effenberg et al (2016) and Effenberg et al (2005) references and added more discussion on the effects of sonification on other cognitive and motor tasks

5- Please specify precisely why the O1, F1, and C3 regions were selected for generating the auditory pattern.

These regions were used because they are the standard electrode locations used for sleep staging in the AASM system–we have added this to the paper.

6- The methodology is well-written, but it should include more details regarding the physical characteristics of the sound (frequency, volume, pitch, timbre, etc.)

We have added these details to the stimulus generation procedure. We have also provided examples of the generated files and corresponding visual EEG traces.

7- It would be better to describe sleep staging precisely and in detail in both the introduction and methodology sections, while also specifying the characteristics of each stage.

We have added more discussion of sleep staging and the characteristics of each stage to these sections

8- The results section is well-reported. The discussion is also good, but it would be better to provide more speculation regarding the reasons behind the obtained results. It would be beneficial to study recent research related to sleep and its influencing variables, and to explain the results with the help of proposed hypotheses. The discussion is currently incomplete and inadequate.

We have revised the discussion section to be clearer about our interpretation of the results–in particular, why we think we observed a benefit for inexperienced novice sleep stagers but not for experienced sleep stagers. However we don’t think it would be helpful to talk about the factors that influence sleep as the focus of this study is primarily on sleep staging as a perceptual learning task, not on sleep itself.

Reviewer 2 comments

Manuscript “EEG Sonification Improves Sleep Staging Performance in Novice Stagers” aimed at investigating whether adding an auditory representation (sonification) of the EEG to a standard visual representation could improve sleep staging performance. Additionally, the study examines if this method reduces workload. The authors found that adding sonification did not result in improvements in accuracy, speed, or workload for the entire subject group .However we found sonification improved accuracy for the least experienced participants. This study explores an interesting and promising concept. Below are some suggestions that should be considered by the authors of the manuscript:

Specific comments

Abstract

1. In the abstract, also mention the classification of groups based on the level of experience.

We have added this to the abstract

Methods

2. How many groups did you have in total? Please provide a detailed explanation of how the groups were divided based on the level of experience and the reception of the Sanifcation sound.

We defined five experience groups based on polysomnograms scored in the last year –these are the groups shown in table 1. We have updated the methods to clarify that these were the groups participants were divided into.

Introduction

1. The first paragraph of the “Manual Sleep Staging” section needs more references.

We have added more references and explanation of the process of sleep staging

2. Hypothesis 1 should be written in full. In which aspect does it perform better?

We have clarified that this is referring to increased agreement with the gold standard scorers as measured by cohen’s kappa

Statistical analysis

1. Why is only the prior year experience of the participants considered? It seems that if the overall experience of the participants is taken into account, the accuracy of the study would be higher.

We chose to measure the prior year as (1) we felt that people would be less precise when estimating the total number of nights staged across their career and (2) recent practice in sleep staging would be the most relevant. We have also added this to the paper.

Discussion

The consequences of the findings in relation to the hypotheses proposed in this study should be clarified with the existing literature.

We have further elaborated on how these results relate to other sonification studies and clarified our explanation of why we think we see results only for novice sleep stagers.

How do you justify the effectiveness of Sanification for novices? You need to reference relevant articles on this.

This is based purely on our observation that novices (people who had scored 10 polysomnograms or less) performed better with sonification than without sonification in our study. We have clarified this in the text.

Paragraph 5 is not relevant to this section. It is suggested that it be briefly mentioned in the Method section.

We have moved this paragraph

Attachment

Submitted filename: Response to reviewers.docx

pone.0329711.s002.docx (18.6KB, docx)

Decision Letter 1

Hesam Ramezanzade

22 Jul 2025

EEG Sonification Improves Sleep Staging Performance in Novice Stagers

PONE-D-25-11016R1

Dear Dr. Whitmore,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice will be generated when your article is formally accepted. Please note, if your institution has a publishing partnership with PLOS and your article meets the relevant criteria, all or part of your publication costs will be covered. Please make sure your user information is up-to-date by logging into Editorial Manager at Editorial Manager® and clicking the ‘Update My Information' link at the top of the page. If you have any questions relating to publication charges, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Hesam Ramezanzade, Ph.D

Academic Editor

PLOS ONE

Additional Editor Comments (optional):

Reviewers' comments:

Acceptance letter

Hesam Ramezanzade

PONE-D-25-11016R1

PLOS ONE

Dear Dr. Whitmore,

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now being handed over to our production team.

At this stage, our production department will prepare your paper for publication. This includes ensuring the following:

* All references, tables, and figures are properly cited

* All relevant supporting information is included in the manuscript submission,

* There are no issues that prevent the paper from being properly typeset

You will receive further instructions from the production team, including instructions on how to review your proof when it is ready. Please keep in mind that we are working through a large volume of accepted articles, so please give us a few days to review your paper and let you know the next and final steps.

Lastly, if your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

You will receive an invoice from PLOS for your publication fee after your manuscript has reached the completed accept phase. If you receive an email requesting payment before acceptance or for any other service, this may be a phishing scheme. Learn how to identify phishing emails and protect your accounts at https://explore.plos.org/phishing.

If we can help with anything else, please email us at customercare@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Hesam Ramezanzade

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    Attachment

    Submitted filename: Re.docx

    pone.0329711.s001.docx (13KB, docx)
    Attachment

    Submitted filename: Response to reviewers.docx

    pone.0329711.s002.docx (18.6KB, docx)

    Data Availability Statement

    Data and code used to perform all the analyses in this study are available at https://doi.org/10.5061/dryad.3bk3j9kz0.


    Articles from PLOS One are provided here courtesy of PLOS

    RESOURCES