Abstract
During wake, sound‐evoked and spontaneous neural activity of the auditory cortex evolves in distinct subspaces, whereas anesthesia disrupts sound responses and merges these spaces. To evaluate if similar modifications of sound representation geometry explain sensory disconnection during sleep, large neural populations of the mouse auditory cortex are followed across slow‐wave sleep and wakefulness. It is observed that sleep dampens sound responses but preserves the geometry of sound representations such that they remain separate from spontaneous activity. Moreover, response dampening is strongly coordinated across neurons and varied throughout sleep, spanning from fully preserved response patterns to population response failures on a fraction of sound presentations. These failures rarely occurred in wakefulness and are more common during high spindle‐band activity. Therefore, in sleep, the auditory system preserves sound feature selectivity up to the cortex for detailed acoustic surveillance but concurrently implements an intermittent gating mechanism leading to local sensory disconnections.
Keywords: auditory perception, NREM sleep, population coding, two‐photon imaging
Sleep is associated with a sensory disconnection whose mechanisms remain elusive. Large neuronal population recordings in the auditory cortex revealed that, in NREM sleep, the neural code for sounds is highly similar to wakefulness, but coordinated modulations of neuron responsiveness intermittently disconnect the local cortical networks from sensory information. This suggests an alternation between surveillance and internal processing in sleep.

1. Introduction
Sleep is characterized by reduced motor activity, decreased behavioral responsiveness, and an absence of memorization of external stimuli, which are often interpreted as a sensory disconnection of the brain from the environment.[ 1 , 2 , 3 ] Several neural mechanisms have been proposed to explain this disconnection, including a gating of sensory information by the thalamus based on its reduced activity, burst mode dynamics, and the occurrence of sleep spindles.[ 4 ] However, cortical sensory responses are still observed during sleep, in particular for its auditory part,[ 5 , 6 ] even during spindles,[ 7 ] which invalidates this gating hypothesis. Moreover, the brain of sleeping subjects still specifically responds to complex stimuli, including the subject's name,[ 8 ] extreme threats,[ 9 ] or learnt behavioral associations.[ 10 ] This remnant processing may preserve some environmental monitoring. At the same time, core sleep processes such as memory consolidation or network homeostasis appear to require isolation from external input to faithfully reflect internal constraints.[ 11 , 12 ] This raises the question of how the brain combines surveillance and offline processes in sleep.
Inspired by the recent observation that, in spite of continued individual neuron responses, population‐level sensory representations are deeply modified under anesthesia,[ 13 ] we reasoned that a specific re‐organization of sensory representation geometry in sleep could subserve these two antagonist functions. We therefore simultaneously followed the activity of up to a thousand neurons across wakefulness and sleep in the mouse auditory cortex during spontaneous activity and sound processing. We observed that sleep significantly reduces the magnitude of population responses to sounds and the amount of available sound information. However, contrary to anesthesia, this reduced responsiveness preserves the structure of sound representation and the separation of spontaneous and evoked activity subspaces. This indicates that sensory processing is intact but dampened. Focusing on single‐trial population responses, we observed that this dampening is due to an adjustment of response gain, coordinated across local cortical populations, which intermittently leads to a full disconnection of the network from the incoming sound input. This intermittent switching between normal sensory processing and disconnection may allow combining auditory monitoring with offline functions.
2. Results
2.1. Robust Sleep During Head‐Fixed Two‐Photon Calcium Imaging
To follow spontaneous and sound‐evoked activity across different sleep states and wakefulness, we trained mice to remain head‐fixed under a fully silent acousto‐optic two‐photon microscope (Figure 1A,B; Figure S1A, Supporting Information). We tested two head‐fixation configurations: on a running‐wheel (12 sessions) with the possibility to locomote or in a small box (13 sessions) on which mice could not perform locomotion behaviors when awake. During the imaging session, we alternated 2 min periods of silence and 2 min periods of sound presentations (60 sounds, ≈ 500 ms duration, inter‐sound interval 2 s) to sample spontaneous and sound‐evoked activity. We imaged layer 2/3 neurons expressing GCAMP6s under the synapsin promoter using simultaneous acquisition of 4 planes at a global sampling rate of 23 Hz (Figure 1C), yielding on average 649 ± 307 neurons per session. Calcium signals were extracted after motion correction[ 14 ] and the calcium event detection algorithm MLSpike was used to denoise calcium signals into an event time series that approximates the actual spike trains of neurons[ 13 , 15 ] (Figure 1D; Figure S2, Supporting Information for extra examples). The sounds included pure tones, frequency chirps, tones with linear and sinusoidal amplitude modulations, broadband noises, and complex natural sounds (Figure 1E). Sleep scoring was performed in an automated manner based on local field potentials recorded in the dorsal hippocampus, the olfactory bulb and auditory cortex, contralateral to the imaging window, and on an electromyogram (EMG) electrode placed on the neck muscle, using previously established methodology[ 16 ] (Figure 1A,F‐I). Wakefulness and sleep were separated based on EMG power and on olfactory bulb gamma band power, which provides a movement‐independent measure of arousal, confirming the identification of sleep in head‐fixed conditions. Both OB gamma and EMG were low during sleep (Figure S1B,C, Supporting Information). REM (Rapid Eye Movement) sleep was identified based on high values of the ratio of theta‐band to delta‐band power in the hippocampus during sleep. During Non‐REM (NREM) periods, we observed, as expected, a large increase of the slow wave and delta‐band power compared to wakefulness and REM in the cortex, confirming the validity of our scoring approach (Figure S1D–F, Supporting Information). Automated sleep scoring was performed in freely moving and head‐fixed conditions in the same animals. After about two weeks of head fixation training of up to four‐hour sessions, we obtained very similar proportions of NREM sleep and wakefulness as in freely moving conditions (Figure 1J), with sleep occurring toward the end of sessions (Figure S1B–D, Supporting Information). In freely moving animals, NREM sleep bouts lasted up to a few minutes, but these longer bouts were rare during head‐fixation (Figure S1G, Supporting Information). REM sleep was also observed under head‐fixation, but about twice less often as in freely moving mice (Figure J). During two‐photon imaging, mice were rarely awoken by sounds, given that only 2.00 ± 0.98% of sound presentations resulted in animals awakening, and sounds marginally affected overall sleep duration (Figure S1H,I, Supporting Information). Awakening probability was not influenced by the type of sound or its intensity (Figure S1J,K, Supporting Information). We also observed that sleep quality improves throughout the session, as deduced from the increased REM sleep duration and the higher sigma band power during the second half of the session (Figure S1L,M, Supporting Information). Our protocol allowed us to identify large epochs of sleep and wakefulness during both silence and sound presentation periods (Figure 1K) during which we could record neural activity of the same population of neurons, visualized using Rastermap[ 17 ] (Figure 1L). As reported previously,[ 13 , 18 ] layer 2/3 auditory cortex spontaneous activity in the awake state was characterized by a mixture of desynchronized activity and population bursts (Figure 1L), and we observed qualitatively the same structure in NREM sleep (Figure 1L).
Figure 1.

Robust sleep during head‐fixed two‐photon calcium imaging. A) Sites of Local Field Potential (LFP) electrodes implantation and AAV‐GCaMP6s injection. A cover slip was placed over the right auditory cortex for imaging. AC = Auditory Cortex, dHpc = dorsal Hippocampus, OB = Olfactory Bulb. B) Recording setup. Mice were allowed to move on a treadmill, while head‐fixed under the objective of a two‐photon microscope. Sounds were played when the animals were awake and asleep. C) Two‐photon Ca2+ simultaneous recording of up to 1200 layer 2/3 neurons expressing GCaMP6s in 4 planes of different depth at 23 Hz. D) Deconvolution of calcium traces with MLSpike. Raw fluorescence signal is fitted with an exponential kernel to infer spike timings. Spike counts are binned in 43 ms windows and smoothed with a 100 ms Gaussian filter. E) Spectrograms of 6 example sound stimuli from different categories: pure tones, chirps, sinusoids, ramps, complex, and white noise. F) LFP example recordings at 1250 Hz of an animal in the awake and the NREM sleep state. G) The evolution of EMG power (top) reflects the arousal state of the animal. The spectrogram (middle) shows the evolution of theta and delta oscillation power changes in the dorsal Hippocampus with sleep stage changes, as the theta/delta ratio (bottom) increase reflects REM sleep. H) The bimodal distribution of EMG power identifies periods of sleep with low power and wakefulness with high power. The bimodal distribution of dHPC theta/delta ratio during sleep unveils periods of NREM (lower values), and REM sleep (higher values). The distribution of these variables outlines the 3 distinct arousal regimes. I) 10 min long extract of the hypnogram resulting from the sleep‐scoring procedure. The animal falls into NREM sleep before entering REM sleep and wakes up. J) Proportion of time spent in each arousal state. Head‐fixation did not change the time proportion spent asleep with regard to the freely‐moving condition, but decreased by half the quantity of REM sleep. K) The recording was separated in 2 min long blocks where either sound stimuli were all presented once (red bars), or no sounds at all were presented (blank). The type of block was selected in a pseudo‐random fashion and independently from the arousal state of the animal. L) Population raster plot of simultaneously recorded neurons during blank (top) and sounds (bottom) blocks, while the animal was awake (left) or asleep (right).
2.2. Reduced Sound Discriminability During NREM Sleep
We first compared sound responses in the auditory cortex across sleep and wakefulness. Given the small amount of REM sleep in imaging sessions, we focused the analysis on NREM sleep and excluded REM periods. Plotting trial‐averaged sound responses of specific neurons, we observed, as previously reported,[ 6 ] that neurons which responded to particular sounds in wakefulness respond to the same sounds in sleep (Figure 2A). These responses, however, had a weaker firing rate (Figure 2A). In some, rarer, cases already weak responses seen in wakefulness almost disappeared in sleep (Figure 2A). Overall, the population firing rate was approximately halved (55.9%) across all sounds in sleep (Figure 2B,C) indicating that despite qualitative preservation of sound responses, the neural activity carrying sound information is dampened in sleep. We estimated the sound‐specific response ratio by dividing all sleep response amplitudes by the response ratio for the sound driving the largest response in wakefulness (Figure 2B). We thereby observed a stronger reduction for simple, artificial sounds than for complex sounds (Figure 2B). The decrease in activity levels during NREM was consistent across the entire population of layer 2/3 neurons, regardless of their average firing rate, indicating a broad, population‐wide shift in neural activity (Figure S3A–D, Supporting Information) and was observed whether or not the mouse could locomote in wakefulness (Figure S3E,F, Supporting Information). This decrease was stronger in NREM bouts of the second half than of the first half of the sessions (Figure S3G, Supporting Information), suggesting a correlation between sleep depth or quality (Figure S1L,M, Supporting Information) and sound response attenuation in the cortex. This reduced firing rate may result in a decrease in available sound information. To quantify this, we used a cross‐validated logistic regression classifier to measure the fraction of single‐trial population responses for which the identity of the presented sound was correctly decoded. We observed a ≈15% decrease in decoder accuracy during NREM (Figure 2D), thus indicating reduced sound information available in local neural populations of the auditory cortex. Misclassification, like the decrease of response gain (Figure 2B), affected all sounds (Figure S4A,B, Supporting Information), was more likely to occur when population responses were weak (Figure S4E,F, Supporting Information), was stronger in the second half of sessions (Figure S4G, Supporting Information), and occurred whether or not the mouse could locomote when awake (Figure S4H,I, Supporting Information). As a point of comparison, we performed the same analysis on previously acquired data across wakefulness and isoflurane anesthesia,[ 13 ] which showed a massive sound information drop (> 3 fold) under anesthesia (Figure 2E). Moreover, whereas sleep confusion matrices had a similar structure as in wakefulness, anesthesia generated systematic classification errors (Figure S4A–D, Supporting Information). This shows that the reduction of sound responses observed in sleep results in a mild loss of sound information and is qualitatively different from the profound disruption observed under anesthesia.
Figure 2.

Conserved but reduced sound discriminability during NREM sleep. A) Averaged sound‐evoked responses of 4 example neurons to each of the 60 stimuli presented (rows) sorted per category. Neurons respond to the same stimuli when the animal is awake or in NREM sleep, except for rare occasions where some responses almost disappear in NREM sleep (white arrows). B) Mean firing rate of neuronal population response relative to baseline is stronger in wakefulness than in NREM sleep for each of the 60 sounds presented. The solid line is the average over the 25 recording sessions, colored area is the standard error of the mean. Rescaling of the NREM curve so that its maximum corresponds to the maximum wake responses (dashed black line) highlights a stronger decrease of responses to all sound categories in NREM sleep relative to complex sounds. C) Average population firing rate relative to baseline over all sounds is stronger in wakefulness than in NREM sleep (Wilcoxon signed rank test, p=3.3 × 10−6, n=25 recordings). D,E) Linear support vector classifiers trained to decode sound identity from single trial neuronal population activity perform above chance level (1.67%, dashed black line) but slightly better in wakefulness than in NREM sleep (D) Wilcoxon signed rank test, p=2.2 × 10−5, n=25 recordings), and accuracy drastically drops isoflurane anesthesia (E) Wilcoxon signed rank test, p=3.1 × 10−2, n=6 recordings).
2.3. NREM Sleep Largely Preserves the Geometry of Sound Representations
Information loss can either be due to a modification of the geometry of representations (i.e., a recomposition of similarity relationships between sounds) or simply to a decrease of signal‐to‐noise ratio. In order to determine which of these two mechanisms underlies information filtering in sleep, we used a noise‐corrected metric of similarity,[ 19 ] insensitive to the norm of activity pattern, to compare the geometry of sound representation across awake and NREM sleep states (Figure 3A). The properties of the chosen metric allow identifying modifications of representation geometry itself independent of modifications of representation amplitude or variability, which relate the signal‐to‐noise ratio. Using this metric, we computed representation similarity analysis (RSA) matrices for the presented sounds, which summarize the representation geometry in the awake and NREM sleep states (Figure 3B; Figure S5A,B, Supporting Information). We observed that RSA matrices were very similar across states (Figure 3B), indicating that despite the small drop in information content, the structure of sound representation geometry in the awake auditory cortex is preserved in NREM sleep. We then computed the noise‐corrected similarity between sleep and wake sound representations to generate cross‐state RSA matrices (Figure 3C) and measured an average similarity level of 0.82 ± 0.09 for the same sounds across the two states (Figure 3D). Thanks to our noise correction, the maximum similarity level of 1 would be reached with a confidence interval of 0.96 to 1.02 if representations were identical (Figure 3D). Hence, while the differences in the representation of sounds in NREM sleep and wakefulness are subtle, they can be statistically detected. We could thereby observe small differences in the level of preservation of the representations depending on several factors. Complex sound representations were slightly better preserved in NREM than simpler sounds (Figure S5E,F, Supporting Information). Deeper sleep appeared to reduce slightly but significantly the similarity of wake and NREM sound representation as estimated from the comparison of the first and second halves of the session (Figure S5G, Supporting Information). Moreover, the similarity of wake and NREM sound representation was larger when mice could not locomote, suggesting that locomotion signal during wake could contribute to the small representational changes (Figure S5H, Supporting Information). As a comparison, the representation similarity across wakefulness and isoflurane anesthesia, based on our previous data,[ 13 ] is much lower, 0.33 ± 0.31 (Figure 3E), than for wakefulness and NREM in any of these different experimental conditions. Moreover, the structure of representation geometry is widely modified by anesthesia[ 13 ] (Figure S5A–D, Supporting Information). Therefore, while anesthesia profoundly disrupts representation geometry, contributing to a large drop in sound information in the auditory cortex, NREM sleep preserves representation geometry and only marginally modifies sound response patterns themselves. This shows that despite the identified effects of NREM sleep on sound representations, they remain largely compatible with awake representations. Therefore, a reorganization of auditory cortex sound representations in sleep is not a mechanism that can explain auditory disconnection.
Figure 3.

NREM sleep largely preserves the geometry of sound representations. A) Schematic explaining the construction of correlation matrices. Neuronal activity is averaged over time between sound presentation onset and 250 ms after offset and concatenated into a vector for each sound. A high correlation value between two sound response vectors indicates a high similarity of neuronal population activation patterns. B) Noise corrected correlation matrices reveal identical similarity relationships between sounds in wakefulness (left) and NREM sleep (right). C) Matrix of pairwise noise corrected correlation between wakefulness and NREM sleep sound response vectors. D) Noise corrected correlation between same sound for wake and NREM sleep response vectors (diagonal of the matrix C) reveals a similarity level largely above chance (Wilcoxon signed rank test, p=1.6 × 10−11, n=60 sounds), but slightly below maximum expected level if representations are identical (95% confidence interval marked as grey area centered around 1). E) Noise corrected correlation between same sound wake and isoflurane induced anesthesia response vectors is slightly above chance (Wilcoxon signed rank test, p=3.1 × 10−12, n=50 sounds).
2.4. Spontaneous and Evoked Activity Cover Different Subspaces in NREM Sleep
It was recently observed that, contrary to the wake state in which spontaneous and sound‐evoked activity are clearly distinct,[ 20 ] under isoflurane and ketamine anesthesia, they produce very similar neural population patterns.[ 13 ] This phenomenon could explain sensory disconnection because if evoked activity is indiscriminable from ongoing activity in the auditory cortex, then it is not possible for downstream target areas to identify if a sound is present or not. To determine if a similar phenomenon occurs in NREM sleep, we collected spontaneous activity from extended periods of silence and sound‐evoked activity patterns in time bins of 43 ms (imaging frame duration) and decomposed them using principal component analysis, separately for wakefulness and NREM sleep. Projections of spontaneous and sound‐evoked patterns in the reduced space of the second and third components qualitatively showed that spontaneous and evoked activity patterns are distributed in different regions of the activity state space both in wakefulness and NREM sleep, but not under isoflurane anesthesia (Figure 4A). This suggests that during sleep spontaneous and evoked activity have little overlap, as in wakefulness. To quantify this further, we measured the cross‐validated amount of sound‐evoked activity variance captured by the spontaneous activity subspace (Figure 4B). The spontaneous subspace was defined as the activity captured by the N first principal components, which accounted for more variance than expected by chance (Figure S6A, Supporting Information, chance is estimated based on a shuffling of cells activity in time in spontaneous activity patterns, see methods). This approach allows us to estimate the dimensionality of the spontaneous activity subspace in the recorded population, and the extent to which evoked activity is included into the spontaneous activity subspace (Figure 4B). The reasoning is that if spontaneous and evoked activity are fully overlapping, the amount of variance captured by the projection of evoked activity onto the spontaneous activity subspace should be equal to the amount of variance captured by the projection of spontaneous activity onto its own subspace. Dimensionality of spontaneous activity was variable across sessions due to variations of neural population size but was large (several 10′s to a few 100′s of principal components) both in the awake and NREM sleep states (e.g., Figure 4C; Figure S6A,B, Supporting Information), while under isoflurane anesthesia, at maximum, the first ≈20 components accounted for most of the significant variance in the spontaneous activity data (Figure 4C; Figure S6A,B, Supporting Information). In wakefulness, as expected from previous observations in the auditory[ 13 ] and visual[ 20 ] cortex, the projection of evoked activity onto the spontaneous activity subspace captured only a small fraction of evoked activity (Figure 4C,D), and only the very first principal components of spontaneous activity significantly captured evoked activity (Figure 4C). Strikingly, the same analysis in NREM sleep also demonstrated a limited overlap between spontaneous and evoked activity (Figure 4C,D), contrary to anesthesia (Figure 4C–E). Notably, spontaneous activity in wakefulness and sleep populated clearly different subspaces (Figure S6C, Supporting Information) in line with the profound change of cortical dynamics in NREM sleep compared to wakefulness, while sound‐evoked activity in NREM and wakefulness shared their subspaces to a greater extent (Figure S6D,E, Supporting Information) as expected from the similarity of population response patterns across these states. The separation of sound‐evoked and spontaneous activity in NREM could be observed in mice recorded in the box and the wheel setup (Figure S6F–H, Supporting Information), ensuring that orthogonality of the subspaces is not just explained by locomotor activity. Moreover, the magnitude of the subspace separation was identical in the second and first halves of the sessions suggesting that sleep depth does not influence the relation between spontaneous and evoked activity (Figure S6I, Supporting Information). Overall, this shows that no gating mechanism based on the similarity of spontaneous and evoked activity is implemented in NREM sleep.
Figure 4.

Spontaneous and evoked activity cover different subspaces in NREM. A) 2D example projection of neuronal activity in principal component (PC) space reveals distinct dimensional occupancy for spontaneous and sound‐evoked activity in wakefulness (left) and NREM sleep (middle), but overlap in isoflurane induced anesthesia (right). B) Schematics explaining the calculation of the fraction of sound‐evoked variance captured by the spontaneous activity space. The sound‐evoked activity is projected into a subspace of the spontaneous activity space defined as the ensemble of significant principal components derived from a first subset of spontaneous activity data. To estimate the level of overlap of spontaneous and sound‐evoked activity, we compared the variance of this projection with the variance of the projection of a second subset of spontaneous activity data that serves as a control for the maximum possible overlap. C) The fraction of variance captured from the sound‐evoked activity projected on the PCs of spontaneous activity is much lower than the maximum overlap expected in both wakefulness (left) and NREM sleep (middle) and this gap increases when adding successive PCs. Under isoflurane anesthesia (right) even the last PCs show a strong overlap between spontaneous and evoked activity. Examples from single recordings. The arrows point toward the values of variance captured by the entirety of the defined reference spontaneous subspace. D, Quantification of fraction of variance captured from all recordings in the different states (Wilcoxon signed rank test, Wakefulness: “Spontaneous reference” to “Projected evoked”: p=6.0 × 10−8, “Projected evoked” to “Chance level”: p=1.2 × 10−7, n=25; NREM sleep: “Spontaneous reference” to “Projected evoked”: p=6.0 × 10−8, “Projected evoked” to “Chance level”: p=6.0 × 10−8, n=25 recordings; Isoflurane induced anesthesia: “Spontaneous reference” to “Projected evoked”: p=6.2 × 10−2, “Projected evoked” to “Chance level”: p=3.1 × 10−2, n=6 recordings). E) The difference in sound‐evoked subspace into spontaneous subspace inclusion level between wakefulness and NREM sleep is smaller than between wakefulness and isoflurane anesthesia (Mann‐Whitney U test, p=3.3 × 10−2). Positive values indicate increased inclusion of sound‐evoked activity in spontaneous subspace compared to wakefulness.
2.5. Intermittent Gating of Local Population Responses During NREM Sleep
Our results so far indicate that auditory cortex sound responses have preserved geometry during sleep and are not masked by spontaneous activity. Thus, the auditory cortex processes sound information with little distortion, apart from a reduction of sound‐related information due to reduced neuronal response amplitudes. In single neurons, this reduction of response amplitude in sleep reflects both a decrease of single trial response probability (i.e., whether or not a neuron responds) and a decrease of the single‐trial response amplitude (i.e., how strongly a neuron responds, Figure 5A,B). However, from a neural population perspective, response variations could happen independently in each neuron or could be coordinated, leading, in the first case, to a mild reduction of the signal‐to‐noise ratio, and in the second case to potentially severe disruptions of information transmission at specific times when all neural responses are globally reduced. To qualitatively evaluate these two alternatives, we plotted single‐trial population responses, sorted according to the norm of the population response vector (Figure 5C). These plots clearly indicated that population responses are more robust in wakefulness than in sleep (Figure 5C), and that fluctuations of population response amplitude are much larger in NREM sleep, with some sound presentations leading to strong population responses, comparable to wake, and others to an apparent absence of response (Figure 5C). To quantify this more precisely, we selected, for the ten most responsive sounds, the 30% trials with the largest population response amplitude, and averaged them to build a template response pattern, representative of well‐driven responses. We then measured the cosine distance between every single trial pattern and the template response pattern to estimate the amount of sound‐related information conserved at the single trial level. While in wakefulness most trials produced response patterns very similar to the template, in sleep this similarity dropped for a large fraction of trials. The drop in similarity was larger during sleep than expected if single‐neuron trial‐to‐trial variability was uncoordinated as evaluated with independent trial shuffling across neurons (e.g., Figure 5D). Cosine similarity was tightly correlated to the baseline‐subtracted population firing rate (Figure S7A–C, Supporting Information) indicating that low cosine similarity corresponds to low responsiveness of the population above spontaneous baseline firing. Moreover, as we observed qualitatively on activity heatmaps (e.g., Figure 5C) a number of single trial responses had a cosine similarity with the template response pattern close to 0 (e.g., Figure 5D) indicating that the corresponding activity pattern is close to orthogonal to the main response axis. Together, these demonstrate that coordinated fluctuations in response strength are larger during sleep, leading to a complete lack of response for certain sound presentations.
Figure 5.

Intermittent gating of population responses during NREM sleep. A) Example of neuronal responses to sounds highlighting the reduced evoked response amplitude in NREM sleep. Thin traces are single trial sound presentations, thick line is the average over trials. B, Average probability of neurons having a non‐zero response to sounds is higher in wakefulness than in NREM sleep (left, Wilcoxon signed rank test, p=5.4 × 10−5, n=25 recordings). Average response amplitude when non‐zero is also higher in wakefulness than in NREM sleep (right, Wilcoxon signed rank test, p=6.7 × 10−3, n=25 recordings). C × Single trial neuron population responses relative to baseline for example sounds reveal events of coordinated reduction of sound‐evoked activity in NREM sleep that are not seen in wakefulness. Trials are sorted by increasing population response intensity. The greyscale bars qualitatively indicate the level of conservation of population responses from completely maintained to absence of response (null). D) Cumulative distribution of similarity between single trial population responses and the corresponding template sound–evoked response in NREM sleep shows more reduced similarity than in wakefulness. Values not significantly different from zero (indicating that the response is essentially orthogonal to the template) are labeled as “non‐responsive” and indicate suppression of sound encoding at the population level (black dots). Dashed lines indicate the similarity values distribution when neuron response intensities are independently shuffled between trials. E, The proportion of trials for which the population is classified as “non‐responsive” is increased in NREM sleep compared to wakefulness (Wilcoxon signed rank test, p=1.2 × 10−4, n=25 recordings). F) Population response amplitude drops during “non‐responsive” events (Wilcoxon signed rank test, wakefulness: p=1.4 × 10−10, n=82; NREM sleep: p=6.5 × 10−23, n=154) and is as low in wake and NREM sleep (Wilcoxon signed rank test, p=8.2 × 10−1, n=69). Amplitude remains higher in wakefulness than NREM sleep when the population is responsive (Wilcoxon, signed rank test, p=2.1 × 10−27, n=230). Inset shows the average sound‐evoked response during “responsive” (top) and “non‐responsive” (bottom) events in NREM sleep. G) Sound decoding accuracy during wakefulness is unchanged when “non‐responsive” events are removed (Wilcoxon signed rank test, p=6.1 × 10−1, n=25 recordings), but slightly increased in NREM sleep (Wilcoxon signed rank test, p=4.9 × 10−4, n=25 recordings). Removing “non‐responsive” events from NREM sleep data was not sufficient to reach the accuracy level of wakefulness data (Wilcoxon signed rank test, p=1.2 × 10−4, n=25 recordings).
To identify these non‐responsive events under statistical control, we used a bootstrap test to evaluate trials with response patterns orthogonal to the direction of well‐driven, template sound response patterns (Figure 5D, see methods). This showed that 17 ± 14% of sound presentations in NREM sleep resulted in an absence of response while this fraction was much smaller in wakefulness (7.8 ± 11%, Figure 5E) and often close to zero in well‐driven wake populations (Figure 5C–E). Non‐responsive events were clearly visible in the raw ΔF/F trace, confirming that our spike inference algorithm did not underestimate responses. Using this signal, we also found a higher proportion of absent responses to sounds during NREM sleep compared to wake (Figure S7D–F, Supporting Information). The increase in null responses in NREM sleep was observed for all sound categories with similar probabilities (Figure S7G,H, Supporting Information), and was stronger in the second half of the session (Figure S7I, Supporting Information) in which sleep appeared deeper (Figure S1L,M, Supporting Information) and did not depend on whether or not the animals had the possibility to have locomotor activity on the setup (Figure S7J, Supporting Information). Trials with these null responses (i.e., orthogonal to the main axis of response) can be seen as non‐responsive trials because their baseline‐corrected population activity is close to 0 (Figure 5F). This indicates no changes relative to ongoing, spontaneous activity and thus a lack of information about the sound. We observed that non‐responsive trials contribute to a fraction of the sound information loss in the auditory cortex during sleep, the remaining fraction being imputable to the other trials with a reduced, but non‐zero response gain compared to wakefulness (Figure 5G).
Finally, we asked if the lack of neural response to sounds we observed in these local neural populations is related to indicators of local or global state in the brain and body. Non‐responsive trials were associated with a weak but significant increase in baseline spontaneous activity amplitude in the imaged neural population just before the sound (Figure 6A), that was not associated with a particular change of the spontaneous activity dynamics as estimated from population activity projections onto its first principal components (Figure 6B). At a more global scale, power in low‐frequency bands, including the delta (2–5 Hz) and the sigma (10–15 Hz, spindle associated) bands was increased in the contralateral auditory cortex around the time of non‐responsive trials in NREM sleep (Figure 6C; Figure S8A–C, Supporting Information). These indicate that the full blockade of sound information observed locally is more probable during oscillatory patterns typically associated with deeper sleep and sensory disconnection.[ 21 , 22 , 23 ] However, non‐responsive trials were not systematically associated with spindle events (Figure 6D; Figure S9, Supporting Information) and spindle detection probability was less than 10% around non‐responsive trials based on automated spindle detection algorithm (Figure 6E). In line with the increase of the sigma band power, more spindles were detected before, during and after sound presentation during non‐responsive trials but always with a low probability (Figure 6E). Non‐responsive events therefore cannot be directly explained by the occurrence of a spindle during the sound presentation, corroborating previous observations that spindles do not affect sound‐driven activity in the auditory cortex.[ 7 ] Non‐responsive trials also did not necessarily systematically co‐occur with delta waves in the contralateral auditory cortex (Figure S10, Supporting Information). We moreover found that the early auditory evoked potential in the contralateral auditory cortex had identical amplitudes whether or not an ipsilateral disconnection was observed in the imaged population (Figure 6F,G). Finally, slight muscle contractions elicited by the sounds were still present during non‐responsive trials (Figure 6H). These observations demonstrate that some auditory signals are preserved at least up to the thalamic input to the cortex during non‐responsive trials and therefore, that the intermittent absence of response observed locally in the auditory cortex does not correspond to a global disconnection of the thalamo‐cortical axis from auditory inputs.
Figure 6.

Population gating of sound responses increases with markers of deeper sleep. A) Averaged neuronal activity in the 500 ms preceding a “non‐responsive” event shows increased activity compared to “responsive” event in NREM sleep (Wilcoxon paired test, p=1.3 × 10−3, n=25 recordings). B, PCA on neuronal activity in the 500 ms preceding sound presentation during sleep. Top left: scree plot. Top right and bottom: 2D example projections show no subspace distinction between activity preceding a “responsive” and a “non‐responsive” event. C) Relative change in oscillation power in a 1.5 s long window centered around sound presentation for “responsive” versus “non‐responsive” events, positive values indicate increases in “non‐responsive” compared to “responsive” events. Delta (2–5 Hz), Theta (5–10 Hz), and Sigma (10–15 Hz) oscillation power in the contralateral AuCx is increased in “non‐responsive” events compared to “responsive” events (Wilcoxon signed rank test, Delta: p=7.3 × 10−5; Theta: p=1.3 × 10−3, Sigma: p=1.3 × 10−4, n=25 recordings). Slow‐wave (0.5–2 Hz), Beta (20–30 Hz), and Gamma (50‐80 Hz) oscillation power, however, are unchanged in “non‐responsive” compared to “responsive” events (Wilcoxon signed rank test, Slow‐wave: p=5.2 × 10−2; Beta: p=9.8 × 10−2; Gamma: p=1.3 × 10−1, n=25 recordings), as well as the ripple bandpower (100–400 Hz) in the dorsal Hippocampus (Wilcoxon signed rank test, p=4.7 × 10−1, n=8), the Olfactory Bulb gamma power (50–70 Hz, Wilcoxon signed rank, p=6.2 × 10−1, n=25), and the filtered EMG power (50–300 Hz, Wilcoxon signed rank test, p=8.0 × 10−2, n=25 recordings). D) LFP recording of the contralateral auditory cortex shows that “non‐responsive” events can be surrounded or not by sleep spindles. Filtered AuCx LFP traces (0.5–70 Hz) from 3 different channels are shown in green. The sigma (10–15 Hz) power of the bottom trace is shown in black with its spectrogram. Arrows indicate a visually detected sleep spindle event. Sound presentation is shown as a grey area. E) Probability to find a spindle event plotted against time from sound onset. Note that the increase of spindle density at sound onset corresponds to a global increase in all power bands due to the evoked potential and not a specific increase in spindle events. F) Averaged LFP around sound presentations for 2 example animals recorded in the contralateral auditory cortex during wakefulness, and “responsive” and “non‐responsive” events in NREM sleep. Raw LFP traces were divided by their averaged value over the entire recording to allow comparison between conditions and recordings. The light grey area shows sound presentation time, the dark grey area is the 50 ms interval after sound onset where a peak in the sound‐evoked potential occurs. G) Mean amplitude of the early sound‐evoked potential in the 50 ms window is comparable between the three conditions (Wilcoxon signed rank test, “wake” to “NREM responsive: p=7.1 × 10−1; “wake” to “NREM non‐responsive”: p=1.6 × 10−1; “NREM responsive” to “NREM non‐responsive”: p=3.3 × 10−1, n=25 recordings). H) Filtered EMG power around sound presentation shows a peak at sound onset in both wakefulness and NREM sleep regardless of the auditory cortex neurons population being responsive or not. Standard deviation is displayed as a colored area.
3. Discussion
In this study, we reveal that sounds are encoded with a highly similar geometrical structure between wake and sleep but with a small but significant drop in accuracy linked to an overall reduction in firing rate (Figure 2). We show that this loss is explained by a mixture of reduced overall activity and coordinated silencing within the neuronal population (Figure 5) which increases both with the depth of sleep (Figures S3G, S4G, S5G, and S7I, Supporting Information) quantified through the proxies of broadband low‐frequency brain oscillations power especially in the delta and sigma rhythms and of higher occurrence of REM bouts (Figure 6; Figure S1L,M, Supporting Information).
Previous work has shown that auditory cortex neurons preserve their sound responses during sleep.[ 5 , 6 ] We extend this finding to show at the population level that the geometry of auditory representations is also conserved between sleep and wake. First, the relations between sounds, as quantified with Representation Similarity Analysis (Figure 3), is conserved in spite of differences in spontaneous activity and ongoing dynamics. Moreover, spontaneous and evoked activity are well separated during sleep, excluding the possibility that sensory disconnection in sleep results from a masking of evoked responses by spontaneous activity (Figure 4). This result extends to the NREM sleep state the recent observation of a clear difference between evoked and spontaneous activity in awake animals,[ 20 ] and strongly contrasts with anesthesia during which evoked activity strongly resembles spontaneous activity, both in the auditory[ 13 , 24 ] and visual[ 25 ] cortex. These observations are consistent with the profound anesthesia‐induced disruptions of auditory representations found even at the earliest stages of auditory processing[ 26 ] and show that modulations of sensory processing in sleep and under anesthesia are mechanistically distinct.
Strikingly, we observed that the preserved geometry of sound representation in layer 2/3 neurons of the auditory cortex is paralleled by an ≈ 50% reduction in firing in response to sounds. This reduction of the neural signal strength leads to a reduction in the accuracy with which sounds can be decoded based on neural population activity during sleep. Quantitatively, the sleep‐induced reduction in cortical responses to sounds that we report is large compared to previous reports of no change,[ 5 , 6 ] or changes restricted to soft sounds[ 27 ] or to late phases of responses.[ 28 , 29 ] These studies were performed with electrophysiology which tends to sample deeper layers whereas our recordings were restricted to layer 2/3. This may account for the difference since a previous electrophysiological study carefully sampling across layers identified intermittent shutdowns of layer 2/3 while layer 5 remained active.[ 30 ] Moreover, whereas electrophysiology shows preserved onset responses and reduced late responses,[ 28 , 29 , 31 ] the slow temporal resolution of calcium imaging pools these periods together, thus emphasizing the net reduction in responsiveness.
Despite the reduced response amplitude, the preserved geometry of sound representations during NREM sleep clearly shows that the auditory cortex continues surveilling the environment based on a processing of acoustic information similar to processing performed wakefulness. This is consistent with the demonstration in human subjects that even the extraction of sophisticated auditory features such as word identity continues to operate during sleep[ 10 ] all the way up to nonconscious preparation of motor responses.[ 32 ] This mechanism could be sense‐specific, as the ear remains physically receptive during sleep and audition allows to monitor distal events. This could confer strong evolutionary advantages since awakening to meaningful stimuli[ 8 , 9 ] mitigates the risks associated with sensory disconnection during sleep. However, isolation of the brain from the environment is thought to be crucial to enable processing that faithfully reflects neural internal constraints such as homeostatic plasticity, memory consolidation or emotional regulation.[ 12 , 33 ] In the auditory cortex, the intermittent coordinated reduction in responsiveness that our study uncovers could be a way for this cortical region to recuperate without needing to be permanently switched off during NREM sleep.
Importantly since spontaneous activity is largely orthogonal to evoked responses during sleep (Figure 4), the interference of sound responses with the internally‐generated sleep‐dependent processes such as memory consolidation[ 34 , 35 ] may be minimized. Our data does not precisely delineate the scale at which the intermittent gating of sensory information occurs in the cortex. We show that at least contralateral thalamic inputs and early cortical potentials are preserved during ipsilaterally observed gating events (Figure 5) indicating that these events are not a full shutdown of auditory processing, and are rather restricted to the cortex. We found that local population gating events are correlated with contralateral oscillatory activity modulations (Figure 6; Figure S8, Supporting Information), albeit with low predictive power (Figure 6E,F). This suggests that auditory population responsiveness could be impacted by slow changes in vigilance state[ 36 , 37 ] instead of specific time‐locked events. On the other hand, since sleep can be accompanied by localized changes in the neural circuit,[ 38 ] our data is also compatible with the hypotheses that local patches of cortex intermittently go offline while others preserve monitoring. Independent of the scale at which this mechanism occurs, it provides a powerful way to combine the need for surveillance and the need for offline processing during sleep.
Our results raise the question of where sensory gating happens since the weak reduction in responses with preserved population geometry that we observe seem insufficient to explain the sensory disconnection that is the hallmark of sleep. Our results support the hypothesis of a downstream, frontal gating.[ 1 , 2 ] Perturbations, whether sensory or artificial using TMS or optogenetics, show that frontal regions typically respond to sensory inputs with an evoked downstate that shuts off inter‐area communication.[ 39 , 40 , 41 ] Therefore, largely preserved sound responses in the auditory cortex may not be passed downstream in the processing hierarchy.
4. Experimental Section
Animals
Seven male C57BL/6 mice were used for experiments, all aged between 8 to 24‐week‐old, and had not undergone any other procedure before. Animals were housed in groups of two to four in controlled temperature and humidity conditions (21–23 °C, 45%–55% humidity), were maintained on a 12 h light‐dark cycle, and had ad libitum access to food, water, and enrichment (wooden logs, cotton bedding, and running wheel). All experiments were conducted during the light phase. All procedures were carried out in accordance with the French Ethical Committees #89 (authorization APAFIS#27040‐2020090316536717 v1).
Surgery
Mice were injected with buprenorphine (Vétergesic, 0,05–0,1 mg kg−1) 30 min prior to surgery. Surgical procedures were carried out using 3.5% isoflurane for induction delivered in an airtight box and were maintained at 1%–2% isoflurane via a mask for the rest of the procedure. After induction, mice were kept on a thermal blanket during the whole procedure and their eyes were protected with Ocrygel (TVM Lab). Lidocaine was injected under the skin of the skull 5 min prior to incision.
For local field potential recordings, coated tungsten (.002″ bare, 0.0040″ coated) wires were chronically implanted intracranially in the olfactory bulb OB (centered on Bregma; A/P: 4.2 mm, M/L: ‐0.5 mm, D/V: 2.25 mm), the dorsal hippocampus dHpc (centered on Bregma; A/P: ‐2.2 mm, M/L: ‐1.7 mm, D/V: 1.3 mm), the primary auditory cortex AC (centered on Bregma; A/P: ‐2.5 mm, M/L: ‐3.75 mm, D/V: 1.25 mm) with a motorized stereotactic arm (StereoDrive, Neurostar, Tubingen, Germany), and at the surface of the cerebellum for ground and references. Silver wires (.005″ bare, 0.0070″ coated) were wrapped around the dorsal neck muscles for electromyography. All electrode wires were pinned to 16 channels EIB (Omnetics, Minneapolis, MN) fixated on the skull of the animal with dental cement (Super‐Bond, Sun Medical, Japan).
For calcium imaging, a 5 mm diameter circular craniotomy was performed above the auditory cortex. Three to six injections of 150 nL of AAV1.Syn.GCaMP6s.WPRE (Vector Core, Philadelphia, PA; 1013 viral particles per ml; used diluted 30x) were done at 30 nL min−1 with pulled glass pipettes at a depth of 500 µm and spaced every 500 µm to cover the large surface of the auditory cortex. The craniotomy was sealed with a circular glass coverslip. The coverslip and head‐post for head‐fixation were fixed to the skull using cyanoacrylate glue (3 m Vetbond) and dental cement (Super‐Bond, Sun Medical, Japan).
After surgery, mice received a 30% glucose solution and Metacam (1 mg kg−1) through subcutaneous injection. Mice had Metacam delivered via their drinking bottle in their home cage for five days starting the day following the surgery. Mice were allowed to recover for a week without any manipulation.
Two‐Photon Imaging
Calcium imaging was performed using a custom two‐photon microscope (AODScope, Karthala Systems, Paris, France) equipped with an acousto‐optic‐deflector scanner combined with a pulsed laser (MaiTai‐DS, SpectraPhysics, Santa Clara, CA) set at 920 nm. A 16x Nikon CFI LWD water dipping objective (MRP07220) was used, which provided a field of view up to 450 × 450µm. For each imaging session, a different area of the window within the auditory cortex was selected for imaging to ensure independence of the recorded neural populations. Images were acquired either at 30 Hz in monoplane mode, or at 23 Hz in multiplane mode with 50 µm depth difference between planes. Laser power reaching the sample was between 80 and 120 mW.
Local Field Potential Recordings
Local field potentials were recorded through an FPGA acquisition board from OpenEphys connected to an RHD 16‐channel recording headstage from Intan plugged into the EIB‐16 embedded on the head of the animal. Signals were high‐pass filtered at 0.1 Hz and recorded at 1250 Hz.
Sleep‐Scoring Procedure
Animal vigilance state was classified by a two‐step automatic algorithm described and previously validated in Bagur et al., 2018.[ 16 ] In the first step, the high‐frequency band (50–300 Hz) power in the EMG or the gamma band (50–70 Hz) power in OB oscillations was used to distinguish wakefulness from sleep. The instantaneous power of the corresponding frequency‐band was computed as the Hilbert transform of the filtered signal and smoothed using a 3 s sliding window. The distribution of these values were fit to a mixture of two Gaussian distributions normalized to each having an area of one, so that the intersection of these two distributions is used as a threshold to separate wakefulness and sleep independently from sleep to wake quantity ratio. Values below the threshold were classified as sleep while values above were classified as wakefulness. In a second step, LFP signal from the dHpc during sleep periods only was filtered in the Delta band (2–5 Hz) and the Theta band (5–10 Hz) and their respective instantaneous power was computed as their Hilbert transform and smoothed using a 2s sliding window. A single normal distribution was fit to the distribution of the ratio of Theta / Delta power. The NREM/REM threshold was placed at the point where the fit explained less than 50% of the data, values below the threshold were classified as NREM sleep while values above were classified as REM sleep. Periods of NREM and REM sleep shorter than 3s were merged into the surrounding periods to avoid artificially short epochs. Awakening frequency was quantified as the number of sleep to wake transitions divided by duration, and sound‐evoked awakening probability was calculated by dividing the number of sound presentations starting in a sleep period that elicited a sleep to wake transition in a 1s window after sound onset by the total number of sound presentations starting in a sleep period.
Experimental Protocol and Auditory Stimulation
After the post‐surgery recovery period, mice had their LFP signals recorded in freely‐moving conditions in a 35 × 18 cm home‐cage with litter and food pellets, for 6 to 8 h during two consecutive days to quantify sleep quantity. Then, mice were habituated to head‐fixation for progressively increasing durations (from 20 min to 3 h) over two weeks, either in a box or on a wheel. The box setup consisted of a 10 × 10 cm box with 5 cm high walls 3D printed with polylactic acid, its floor was covered with litter to absorb urine and feces. The wheel setup consisted of a 15 cm diameter horizontal running wheel to allow free locomotion of the animal and was also covered with litter to absorb urine and feces. Each mouse was habituated to a single type of setup only. LFP signals were recorded during habituation sessions.
When an animal was sufficiently habituated to head‐fixation to fall asleep and display REM sleep with a stable sleep architecture over at least two days, it was used for imaging sessions (e.g., Figure S1A, Supporting Information). During the sessions, the mouse was head‐fixed on the setup it was habituated to. The activity of neurons in the primary auditory cortex was recorded with two‐photon calcium imaging through the cranial window, and LFP signals were recorded to monitor the vigilance state of the animal. The recording was split into 80 to 100 blocks of 2 min with either silence or sounds. Equal numbers of silence and sound stimulation blocks alternated randomly to sample spontaneous and sound‐evoked activity during wake and sleep throughout the session. Within a stimulation block, all sound stimuli were presented once in a randomized order. Each sound was 450 to 500 ms long, and sound onsets were separated by a 2 s interval within the block. Sounds were delivered at 192 kHz with an NI‐PCI‐6221 card (National Instrument) driven by a custom MATLAB software through an amplifier and high‐frequency loudspeakers (SA1 and MF1‐S, Tucker‐Davis Technologies). Sounds were calibrated in intensity at the location of the mouse ear using a probe microphone (Bruel & Kjaer).
The sound stimuli consisted of a collection of 60 predefined sounds divided into six groups (Figure 1E): four pure tones (6, 8, 12, and 25 kHz), four frequency‐modulated sounds (4–16, 4–25 kHz, upward and downward modulations), four ramping amplitude modulation sounds (8 and 12 kHz), four sinusoidal amplitude modulation sounds (sinusoidal modulation at 20 and 80 Hz, for two carrier frequencies of 8 kHz and 12 kHz), ten complex sounds (5 recordings of 4× accelerated music, animal calls, and natural environments, high‐pass filtered at 2 kHz, none of these sounds contained mouse calls or other ecologically relevant sounds), and white noises and derivatives (raw white noise, ramping up and down, filtered in the 6.4 to 19 kHZ and 8.6 to 12 kHz band, filtered ≈ 12 and 25 kHz and ramping up and down, and filtered in the 13 to 24 kHz band with amplitude sinusoidal modulation at 40 Hz). All sounds (except for sinusoidal and ramping sounds) were played at “low” intensity (50 or 60 dB SPL) and “high” intensity (70 or 80 dB SPL). Each sound was repeated 40 or 50 times throughout the whole protocol, falling randomly during sleep or wake periods. This protocol ensured an average of 18.8 +/‐ 08.5 presentations per sound during sleep, and 21.7+/‐ 10.9 during wake.
Calcium Signal Processing
Data analysis was performed using Python and MATLAB scripts. Artifacts related to horizontal movements were corrected frame by frame using rigid and non‐rigid body registration. ROI were identified by applying Singular Value Decomposition (SVD) on the movie and finding peaks in the components, and were then curated manually. Registration and ROI identification were performed using the Suite2P algorithm.[ 14 ] Neuropil contamination was subtracted by applying the following equation: Fcorrected (t) = Fmeasured (t) − 0.7Fneuropil (t) to each ROI, where F is the fluorescence and t is time. These corrected traces were deconvolved using the MLSpike algorithm[ 15 ] to estimate the putative timing of spike trains underlying the recorded fluorescence. A time constant of calcium transient τ = 2s in accordance with the known decay time of GCamp6s,[ 42 ] and polynomial coefficients [p2,p3] = [0.73, ‐0.05] for the supralinear calcium indicator response as previously calibrated with patch recordings.[ 15 ] Unitary spike coefficient a = 0.1 and coefficient adjusting for baseline drift compensation (0.2) were estimated by visual inspection over multiple sessions to best fit the descending slope of experimental calcium spikes as well as the fluctuation of the baseline fluorescence. The number of spikes was then counted for each neuron within time bins corresponding to frames acquired by the microscope (33 ms for monoplane acquisition, 43 ms for multiplane acquisition), and this spike count was smoothed by convolution with a Gaussian kernel (std = 100 ms, if not stated otherwise).
Stimulus Decoding Classifiers
Linear Support Vector Classifiers (LinearSVC) were trained to decode stimulus from neural activity. For each recording session, classifiers were fed with the population responses vectors to sound stimuli. Population response vectors were computed by concatenating the response of every neuron recorded in the session to each sound presented during the experiment. Responses were averaged in time over stimulus presentation and the following 250 ms to take into account responses to sound offset. For each session, classifiers were trained with a stratified 5‐fold cross‐validation procedure: the dataset was randomly split in 5 non overlapping groups with the same number of stimuli and a comparable proportion of each stimulus type. To avoid class imbalance bias, the number of presentations per sound was resampled to ensure training used the same number of presentations per sound. The classifier was trained on 4 of the groups and accuracy was tested on the remaining group. The procedure was repeated so that each group plays the role of the test set once. The accuracy of the classifier was computed as the weighted F1 score across stimuli in the test sets. The number of sound presentations were equalized between wake and NREM sleep conditions to ensure that observed differences in performance of the classifiers would not depend on the size of the training dataset between the two conditions.
Representation Similarity Analysis
To assess the level of similarity between neuronal patterns of responses to different sound stimuli, an estimation of the Pearson correlation coefficient correcting for trial‐to‐trial variability was implemented. The level of similarity between two neuronal patterns was computed as: , where 𝜌 denotes the Pearson correlation coefficient, <Si>r is the population vector in response to sound i averaged over trials from the ensemble r. The ensembles r and r’ randomly split the dataset in two halves with equal number of occurrences of each stimulus type. This metric allows to normalize the correlation level to the maximum correlation level that can be expected given the noise in single trial responses. Because responses during wake and NREM sleep do not share the same ensemble of trials, the similarity level of the responses between the two states can be written as: , where r and r’ split in half sound presentations occurring during wakefulness periods, and q and q’ during NREM sleep periods. Similarity of coding was assessed when i = j, and its corresponding chance level was calculated by shuffling sound identity so that it is mismatched between the two states. The expected value obtained when representations are similar is computed by splitting wakefulness data in 4 non‐overlapping groups of sound presentations r, r’, q, q’, and should be centered ≈ 1.
Dimensionality of Neuronal Activity Subspaces
To assess the dimensionality of neuronal activity subspaces PCA was performed on the time series of simultaneously recorded neurons during a specific condition (spontaneous activity recorded during blocks of silence, sound‐evoked activity during sound presentation time, in wakefulness, NREM sleep, and isoflurane induced anesthesia). The time series were z‐scored over the whole recording session. In order to estimate the dimensionality of each activity space, the number of principal components that explained more variance than chance was estimated.
The time series were split into 5 equal blocks of continuous activity for cross‐validation: principal components were estimated using 4 of the blocks (train set) and the remaining block (test set) was projected onto these components to compute the variance they capture: Var(Xtest Vtrain Vtrain T ), with Xtest the time series from the test set, and Vtrain the component matrix of the train set. The procedure was repeated 5 times so that each set was used as the test set once, and the average variance ratio captured in the test sets per component was defined as the cross‐validated PC scree plot. A chance level was computed in the same fashion except that the test set was circularly shuffled 20 times in the time dimension independently for each neuron to break temporal relationships between neurons. The 95th percentile of this new variance ratio captured per component distribution was kept as the chance level. The dimensionality of an activity space was defined as the number of components for which the captured variance remains above this chance level, as visualized using scree plots (Figure S6A, Supporting Information).
Spontaneous and Sound‐Evoked Activity Subspaces Comparison
To determine whether spontaneous and sound‐evoked activity shared the same dimensions an inclusion index of sound‐evoked activity in spontaneous subspace was computed for each recording session. Intuitively, the inclusion index captures how much of the sound‐evoked activity variance was captured by the principal components derived from the spontaneous activity. The spontaneous activity was split into 10 blocks of contiguous activity with equal duration. Blocks with an odd‐numbered label were used as the reference set for spontaneous activity. Dimensions of this reference spontaneous subspace were defined as the ensemble of its first n principal components Vref , n being the previously computed dimensionality of that subspace. The percentage of variance in another time series Xsource that can be explained by dimensions of the reference subspace is with Vsource the matrix of the first n principal components of Xsource . The maximum level of subspace inclusion to be expected %Varmax is obtained when Xsource is the other half of spontaneous activity data. The level of inclusion of sound‐evoked activity in the spontaneous activity subspace %Varprojected is obtained when Xsource is the time series of neuronal activity during sound stimulus presentation (sound‐evoked activity). The chance level %Varchance is calculated with Xsource being the sound‐evoked activity circularly shuffled in time independently for each neuron. The inclusion index was computed as .
Non‐Responsive Events Labeling Procedure
To identify sound presentation trials when the neuronal population is not responsive to sounds, cosine similarity was computed of the single‐trial sound‐evoked response vectors with a template vector computed as the average of the 30% strongest responses to the corresponding sound stimulus. The variability of that similarity estimator with a bootstrapping procedure was estimated: neurons constituting the population were randomly resampled 1000 times with replacement. Trial was labeled as “non‐responsive” if 0 belonged in between the 5th and the 95th percentile of the bootstrapped cosine similarity distribution. The analysis was restricted to the 10 sounds driving the strongest responses in the neuronal population for each recording.
Automatic Detection of Spindles
To identify spindle events, an automatic algorithm was implemented in line with previous studies.[ 7 , 43 ] The AuCx LFP channel with the strongest sigma power during NREM sleep was selected. The LFP trace was filtered in the Sigma band (10–15 Hz) with a zero‐phase forward and reverse digital Infinite Impulse Response filtering, and restricted it to NREM sleep periods. Its instantaneous amplitude was extracted as the element‐wise square of the filtered trace. Events of spindles were considered when the amplitude exceeded 2 times its standard deviation above the mean. Events separated by less than 200 ms were merged together, and events shorter than 400 ms or longer than 3 s were discarded.
Statistical Analysis
All quantification and statistical analyses were performed with custom Python scripts. Statistical assessment was based on two‐sided non‐parametric tests (Wilcoxon signed‐rank, Mann‐Whitney U, Kruskall‐Wallis) reported in the text or in figure legends together with the number of samples used for the test and the nature of the sample (number of neurons, recording sessions or mice). Unless otherwise mentioned, false positive rates below 0.05 were considered significant. Data are presented as box plots. The box extends from the first quartile to the third quartile of the data, with a line at the median. The whiskers extend from the box to the farthest data point lying within 1.5x the inter‐quartile range from the box. Flier points are those past the end of the whiskers. No animal or recording session was excluded from analyses, except for the comparison of neural responses between two session halves, for which one session was excluded because it was too short (<2 h).
Declaration of Generative AI and AI‐Assisted Technologies
The authors did not use generative AI and AI‐assisted technologies at any step of the study and in the writing process.
Conflict of Interest
The authors declare no conflict of interest.
Author Contributions
S.B. and B.B. contributed equally to this work as co‐last authors. A.M., B.B., S.B. performed conceptualization. A.M., S.B., B.B., A.F. performed methodology. A.M., S.B. performed investigation. A.M. performed data curation. A.M. performed formal analysis. B.B. and S.B. performed supervision. B.B. performed project administration. B.B. and S.B. acquired funding acquisition. A.M. performed software. A.M. performed visualization. B.B. and S.B. performed validation. B.B., S.B., A.M. wrote the original draft, wrote reviewed and edited the final manuscript.
Supporting information
Supporting Information
Acknowledgements
The authors thank Dr. Karim Benchenane for his help in the implementation of delta wave detection. The authors acknowledge the support of the Fondation pour l'Audition to the Institut de l'Audition. The authors acknowledge the support of the following funding sources: Fondation pour l'Audition, RD‐2023‐1 (BB), FPA IDA02 (BB), and APA 2016‐03 (BB), European Research Council, ERC CoG 770841 DEEPEN (BB) Fondation pour la Recherche Médicale SPF202005011970 (SB). Agence Nationale pour la Recherche, AudioDream (BB, SB), France 2030 program, ANR‐23‐IAHU‐0003 (BB).
Muller A., Filipchuk A., Bagur S., and Bathellier B., “The Geometry of Layer 2/3 Cortical Sound Processing in Slow Wave Sleep.” Adv. Sci. 13, no. 7 (2026): e09707. 10.1002/advs.202509707
Data Availability Statement
All datasets and custom codes are freely available on Zenodo. doi:10.5281/zenodo.15530880
References
- 1. Andrillon T., Kouider S., Curr. Opin. Physiol. 2020, 15, 47. [Google Scholar]
- 2. Cirelli C., Tononi G., Clin. Transl. Neurosci. 2024, 8, 9. [Google Scholar]
- 3. Coenen A., J. Sleep Res. 2024, 33, 14152. [DOI] [PubMed] [Google Scholar]
- 4. McCormick D. A., Bal T., Curr. Opin. Neurobiol. 1994, 4, 550. [DOI] [PubMed] [Google Scholar]
- 5. Issa E. B., Wang X., J. Neurosci. 2008, 28, 14467. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6. Nir Y., Vyazovskiy V. V., Cirelli C., Banks M. I., Tononi G., Cereb. Cortex 2015, 25, 1362. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7. Sela Y., Vyazovskiy V. V., Cirelli C., Tononi G., Nir Y., Sleep 2016, 39, 1069. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8. Oswald I., Taylor A. M., Treisman M., Brain J. Neurol. 1960, 83, 440. [DOI] [PubMed] [Google Scholar]
- 9. McKenna J. T., Thankachan S., Uygun D. S., Shukla C., McNally J. M., Schiffino F. L., Cordeira J., Katsuki F., Zant J. C., Gamble M. C., Deisseroth K., McCarley R. W., Brown R. E., Strecker R. E., Basheer R., Curr. Biol. 2020, 30, 2379. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10. Türker B., Musat E. M., Chabani E., Fonteix‐Galet A., Maranci J.‐B., Wattiez N., Pouget P., Sitt J., Naccache L., Arnulf I., Oudiette D., Nat. Neurosci. 2023, 26, 1981. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11. Born J., Rasch B., Gais S., The Neuroscientist 2006, 12, 410. [DOI] [PubMed] [Google Scholar]
- 12. Tononi G., Cirelli C., Neuron 2014, 81, 12. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13. Filipchuk A., Schwenkgrub J., Destexhe A., Bathellier B., Nat. Neurosci. 2022, 25, 1327. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14. Pachitariu M., Stringer C., Dipoppa M., Schröder S., Rossi L. F., Dalgleish H., Carandini M., Harris K. D., bioRxiv 2016, 10.1101/061507v2. [DOI] [Google Scholar]
- 15. Deneux T., Kaszas A., Szalay G., Katona G., Lakner T., Grinvald A., Rózsa B., Vanzetta I., Nat. Commun. 2016, 7, 12190. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16. Bagur S., Lacroix M. M., de Lavilléon G., Lefort J. M., Geoffroy H., Benchenane K., PLoS Biol. 2018, 16, 2005458. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17. Stringer C., Zhong L., Syeda A., Du F., Kesa M., Pachitariu M., Nat. Neurosci. 2025, 28, 201. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18. Harris K. D., Bartho P., Chadderton P., Curto C., De La Rocha J., Hollender L., Itskov V., Luczak A., Marguet S. L., Renart A., Sakata S., Hear. Res. 2011, 271, 37. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19. Bagur S., Bourg J., Kempf A., Tarpin T., Bergaoui K., Guo Y., Ceballo S., Schwenkgrub J., Verdier A., Puel J. L., Bourien J., Bathellier B., Sci. Adv. 2025, 11, adr6214. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20. Stringer C., Pachitariu M., Steinmetz N., Reddy C. B., Carandini M., Harris K. D., Science 2019, 364, aav7893. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21. Rechtschaffen A., Hauri P., Zeitlin M., Percept. Mot. Skills 1966, 22, 927. [DOI] [PubMed] [Google Scholar]
- 22. Dang‐Vu T. T., McKinney S. M., Buxton O. M., Solet J. M., Ellenbogen J. M., Curr. Biol. 2010, 20, R626. [DOI] [PubMed] [Google Scholar]
- 23. Blanco‐Duque C., Bond S. A., Krone L. B., Dufour J.‐P., Gillen E. C. P., Purple R. J., Kahn M. C., Bannerman D. M., Mann E. O., Achermann P., Olbrich E., Vyazovskiy V. V., Sci. Adv. 2024, 10, adn6247. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24. Luczak A., Bartho P., Harris K. D., Neuron 2009, 62, 413. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25. Kenet T., Bibitchkov D., Tsodyks M., Grinvald A., Arieli A., Nature 2003, 425, 954. [DOI] [PubMed] [Google Scholar]
- 26. Gosselin E., Bagur S., Bathellier B., Sci. Adv. 2024, 10, ado2291. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27. Issa E. B., Wang X., J. Neurosci. 2011, 31, 2965. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28. Marmelshtein A., Eckerling A., Hadad B., Ben‐Eliyahu S., Nir Y., Curr. Biol. CB 2023, 33, 2925. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29. Sela Y., Krom A. J., Bergman L., Regev N., Nir Y., J. Neurosci. 2020, 40, 2895. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30. Sakata S., Harris K. D., Neuron 2009, 64, 404. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 31. Hayat H., Marmelshtein A., Krom A. J., Sela Y., Tankus A., Strauss I., Fahoum F., Fried I., Nir Y., Nat. Neurosci. 2022, 25, 935. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 32. Kouider S., Andrillon T., Barbosa L. S., Goupil L., Bekinschtein T. A., Curr. Biol. CB 2014, 24, 2208. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 33. Goldstein A. N., Walker M. P., Annu. Rev. Clin. Psychol. 2014, 10, 679. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 34. Bendor D., Wilson M. A., Nat. Neurosci. 2012, 15, 1439. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 35. Salgado‐Puga K., Rothschild G., BioRxiv 2023, 22, 568283, 10.1101/2023.11.22.568283. [DOI] [Google Scholar]
- 36. Lecci S., Fernandez L. M. J., Weber F. D., Cardis R., Chatton J.‐Y., Born J., Lüthi A., Sci. Adv. 2017, 3, 1602026. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37. Cardis R., Lecci S., Fernandez L. M., Osorio‐Forero A., Chu Sin Chung P., Fulda S., Decosterd I., Lüthi A., eLife 2021, 10, 65835. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 38. Andrillon T., Oudiette D., Neurosci. Biobehav. Rev. 2023, 155, 105465. [DOI] [PubMed] [Google Scholar]
- 39. Wilf M., Ramot M., Furman‐Haran E., Arzi A., Levkovitz Y., Malach R., PLoS One 2016, 11, 0157143. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 40. Massimini M., Ferrarelli F., Huber R., Esser S. K., Singh H., Tononi G., Science 2005, 309, 2228. [DOI] [PubMed] [Google Scholar]
- 41. Li B., Ma C., Huang Y.‐A., Ding X., Silverman D., Chen C., Darmohray D., Lu L., Liu S., Montaldo G., Urban A., Dan Y., Cell 2023, 186, 5739. [DOI] [PubMed] [Google Scholar]
- 42. Chen T. W., Wardill T. J., Sun Y., Pulver S. R., Renninger S. L., Baohan A., Schreiter E. R., Kerr R. A., Orger M. B., Jayaraman V., Looger L. L., Svoboda K., Kim D. S., Nature 2013, 499, 295. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 43. Blanco‐Duque C., Bond S. A., Krone L. B., Dufour J. P., Gillen E. C. P., Purple R. J., Kahn M. C., Bannerman D. M., Mann E. O., Achermann P., Olbrich E., Vyazovskiy V. V., Sci Adv. 2024, 10, eadn6247. [DOI] [PMC free article] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Supporting Information
Data Availability Statement
All datasets and custom codes are freely available on Zenodo. doi:10.5281/zenodo.15530880
