Skip to main content
PLOS One logoLink to PLOS One
. 2020 Aug 10;15(8):e0236800. doi: 10.1371/journal.pone.0236800

Enhancement of visual biological motion recognition in early-deaf adults: Functional and behavioral correlates

Marie Simon 1,*, Latifa Lazzouni 1, Emma Campbell 1, Audrey Delcenserie 1,2, Alexandria Muise-Hennessey 3, Aaron J Newman 3, François Champoux 2,4, Franco Lepore 1
Editor: Paul Hinckley Delano5
PMCID: PMC7416928  PMID: 32776962

Abstract

Deafness leads to brain modifications that are generally associated with a cross-modal activity of the auditory cortex, particularly for visual stimulations. In the present study, we explore the cortical processing of biological motion that conveyed either non-communicative (pantomimes) or communicative (emblems) information, in early-deaf and hearing individuals, using fMRI analyses. Behaviorally, deaf individuals showed an advantage in detecting communicative gestures relative to hearing individuals. Deaf individuals also showed significantly greater activation in the superior temporal cortex (including the planum temporale and primary auditory cortex) than hearing individuals. The activation levels in this region were correlated with deaf individuals’ response times. This study provides neural and behavioral evidence that cross-modal plasticity leads to functional advantages in the processing of biological motion following lifelong auditory deprivation.

Introduction

An increasing number of studies suggest that early sensory loss leads to the enhancement of the other intact sensory modalities [1]. Several behavioral studies have shown that early-deaf people possess enhanced abilities for visual localization and visual motion detection [2]. According to functional neuroimaging studies, the visual enhancements in early-deaf individuals are generally attributed to the recruitment of the deafferented auditory cortex [36]. Therefore, the visual crossmodal activity of the auditory cortex is typically defined as compensatory, meaning that deaf people rely more on their intact visual system to encode their environment in comparison to hearing individuals [7]. Some tactile [811] and language abilities (i.e., sign language and/or lip-reading) [1217] are also associated with the recruitment of the auditory cortex in deaf people [1] and support the compensatory reorganization of the brain after early auditory deprivation.

This study’s aim is to tackle the relevant topic of visual crossmodal plasticity following early auditory deprivation with the visual ability to perceive biological motion i.e. gesture sequences that characterize all living things [18]. The study of biological motion is an interesting issue since with only minimal pieces of visual information, such as point-lights at the main joints of the human body, people can efficiently recognize human actions [18,19]. Human movement recognition is essential for social cognition and interaction. With this ability, people can understand the gestural intentions of others and respond adequately [20]. For the deaf individuals using sign language, the adequate comprehension of human action is specifically critical to rapidly detect the presence of linguistic movements [21]. More generally, the ability to quickly recognize human motion also represents additional visual cues for deaf individuals to interpret their environment despite the auditory deprivation [22].

Originally, the cerebral network associated to the understanding of action (biological movement, human action) was referred to as the mirror neurons system [23]. The human mirror neurons system is formed by the inferior parietal lobule (IPL), the ventral premotor cortex (PMv) as well as the inferior frontal gyrus (IFG, BA 44/45) in the homologous brain of the macaque [24]. Henceforth, it is commonly accepted that the cerebral network of action understanding in humans is broader than the previously cited regions and also includes the posterior superior temporal sulcus (pSTS), the supplementary motor area (SMA, BA 6), the primary somatosensory cortex (S1, BA 1/2), the intraparietal cortex (IPS), the posterior middle temporal gyrus (pMTG) at the transition to visual area V5, and fusiform face area/fusiform body area (FFA/FBA) [25]. It is interesting that the neural responses associated to point-light biological motion recognition involve the same characteristic set of regions implicated in human action recognition [26].

In prior studies, several stimuli have been used to disentangle cerebral networks involved in either or both sign language and human action recognition processes between deaf native signers and hearing individuals. Among the human actions, meaningless gestures, pantomimes, emblems, and signs are conceptualized as a continuum in terms of linguistic properties, conventionalization, and semiotics characteristics [27]. Pantomimes are non-communicative gestures that are oriented towards an object, an action or an event [28] who can convey meaning on their own without speech [27]. Emblems are conventional communicative gestures [27] that are culturally influenced [29] and defined as non-verbal action used to convey information to others [30] (for illustration see Fig 1). These two types of gesture are not language per se. They differ from sign languages since the latter are natural human languages that have evolved spontaneously in Deaf communities, and possess all of the linguistic structural properties and complexity of spoken languages [31]. Although sign languages use the visual-manual rather than aural-oral modalities, the networks of brain regions recruited for spoken and signed language processing are largely overlapping [32].

Fig 1. Stimuli and behavioural results.

Fig 1

(A) example of a communicative gesture/emblems « calm down » (B) example of a non-communicative gesture/pantomimes « playing guitar » (C) example of a scrambled versions (D) Behavioral results illustrating the reaction times (RT) according to both groups. Errors bars denote standard deviation.

To date, all of the previous studies on deaf signers fail to converge neuroimaging with behavioral results. Using fMRI, two studies have investigated the cerebral network involved in the passive observation of pantomimes by deaf native signers. These studies report a hypoactivation of the human mirror neuron system in the IFG, and the IPL in deaf signers individuals [21,33]. On the other hand, some neuroimaging studies with pantomimes [34], sequences of meaningless gestures [35,36] or a single emblem [37,38] support a similar human action network between the deaf signers and hearing individuals. In the current study, we attempt to replicate and extend these findings to multiple emblems. This way, the present study offers a robust comparison of the human action network between emblems and pantomimes. Indeed, these two stimuli differ according to whether they aim to transmit information or not, since emblems represent communicative gestures whereas pantomimes represent non-communicative gestures [19]. Additionally, only the emblems show some linguistic properties, such as phonological and morphological components [27]. Furthermore, activation of the superior temporal gyrus (STG), including the primary auditory cortex and the planum temporale, has been observed across tasks requiring to recognize emblems, pantomimes, and meaningless gestures [14,34,37] despite the absence of behavioral differences in terms of accuracy or reaction time between deaf signers and hearing individuals. Together, these findings suggest that lifelong deafness and/or sign language use could lead to alterations in the neural networks recruited to interpret manual communication, even when it is not linguistically structured. Furthermore, increased recruitment of traditionally auditory and language processing areas during gesture recognition may reflect that lifelong reliance on visual communication (sign language and lip-reading) [39] leads to alternative neural strategies for the processing of this information. Moreover, none of the prior studies have included early-deaf people who are not signers but used rather spoken language and explore the distinct effect of linguistic experience and auditory deprivation on visual crossmodal plasticity. The goal of the present study was to compare neural responses to both emblems and pantomimes between early-deaf and hearing individuals, and for the first time to relate these to behavioral performance. Given the lack of convergence in previous studies, we expected that combined behavioral and fMRI results might seize compensatory brain plasticity in early-deaf individuals, independently of their primary mean of communication. To test our hypothesis, we measured the fMRI bold response to emblems and pantomimes recognition in early-deaf individuals who used or not sign language in comparison to hearing peers.

Methods

Participants

Thirty-five French-speaking adults participated in the present study. All the participants provided written informed consent prior to testing and all experiments were performed in accordance with relevant guidelines and regulations. This study was approved by the ethics committee and scientific boards of the Centre de Recherche Interdisciplinaire en Réadaptation du Montréal métropolitain (CRIR) and the Quebec Bio-Imaging Network (QBIN). One deaf and two hearing participants were excluded from the study due to technical problems during fMRI data acquisition. A total of 32 participants were therefore included in the study: 16 early severe-to-profound deaf subjects (11 women, Mean age ± SD = 30.25 ± 4.69 years) were compared to 16 hearing participants (12 women, Mean age ± SD = 30.31 ± 5.42 years) matched on age, sex, and number of years of education. All subjects had a normal or corrected-to-normal vision and no history of neurological pathology. According to the Edinburgh handedness inventory index [40], five deaf and three hearing participants were left-handed. All participants were administered the Matrix Reasoning subtest of the Weschler Abbreviated Scale of Intelligence (WASI-II) [41], which is a brief evaluation of non-verbal intelligence, namely of nonverbal fluid reasoning [42]. The results showed that both groups performed in the average to the superior level of ability, as indicated by T scores (deaf participants: M ± SD = 57.44 ± 4.85; hearing participants: M ± SD = 62.46 ± 4.45).

Deaf participants had a severe-to-profound hearing loss greater than 77 dB HL (M ± SD = 94.11 ± 9.93) in both ears as determined by certified audiologists. Specifically, 13 participants had a hearing loss greater than 90 dB HL at 500, 1000, 2000, 4000, and 8000 Hz in both ears while two participants were able to detect 500 Hz pure tones presented at 80 dB HL and 77 dB HL in their better ear. Four participants reported having hereditary congenital deafness whereas, for twelve participants, congenital or early deafness was due to unknown etiologies. Eight of the sixteen deaf participants were proficient signers and four of them were native deaf signers in the Langue des Signes Québécoise (LSQ). Eight participants had been using hearing aids since childhood, used spoken language only for expression and relied on lip-reading for reception (see Table 1 for detailed information about the participants).

Table 1. Demographic and clinical data for the 16 deaf participants.

Subject Sex Etiology Age Hearing aid Hearing loss: left ear/right ear (dBHL) Primary language WASI T-score Handedness
1 M Unknown 36 No 100/100 Sign 54 R
2 F Genetic 22 No 105/110 Sign (native) 63 R
3 F Genetic 25 No 90/90 Sign (native) 52 L
4 M Unknown 36 No 90/90 Sign 58 L
5 F Unknown 29 Yes 115/110 Spoken 68 R
6 F Genetic 25 Yes 93/95 Spoken 58 L
7 F Unknown 29 No 90/90 Sign (native) 62 R
8 M Unknown 25 Yes 87/92 Spoken 50 R
9 F Unknown 33 Yes 103/102 Spoken 58 L
10 F Unknown 34 Yes 106/106 Spoken 60 R
11 F Genetic 36 No 90/90 Sign 52 R
12 F Unknown 28 Yes 93/92 Spoken 62 R
13 F Unknown 37 Yes 78/77 Spoken 58 R
14 F Unknown 28 No 97/95 Sign (native) 58 R
15 M Unknown 31 No 90/90 Sign 60 L
16 M Unknown 30 Yes 101/106 Spoken 49 R

Stimuli and experimental protocol

The stimuli consisted of 126 point-light animated videos representing 42 emblems (e.g. “calm down”), 42 pantomimes (e.g. “playing guitar”), and 42 scrambled versions of these biological motions (Fig 1). We carefully controlled point-light stimuli, which allowed us to isolate the effects of biological motion from possible confounding effects such as face and body perception. Point-light also allows us to isolate biological motion processing from more general visual motion perception, by including a control condition in which the starting positions of the points are randomized, but their motion vectors remain the same [19]. Previous studies that used videos and often compared gesture conditions to non-motion control conditions, were thus limited in the interpretation of their results [43].

An event-related fMRI protocol was split in two runs both presented in random order across participants. The stimulation task was implemented on Psychopy with Python 3.4. Each run of six-minutes comprised 63 different videos (21 stimuli of each category). Safety instructions and imaging sequences were explained to the participants to familiarize them with the fMRI environment. The participants all performed a training trial of the biological motion task before the fMRI session. The instructions were presented before each run and the participants had to press a button once they were done reading them. Each video lasted between two to four seconds and was followed by an inter-stimulus interval randomly varying from two to ten seconds. Biological motion stimuli were projected on a screen at the back of the scanner and were presented to the participants through a mirror attached to the MRI head coil–at approximately 12 cm away from the eyes. With an fMRI-safe button response pad, participants were asked to press as fast and as accurately as possible with the correct button (1: whether the video was a human motion with no communicative content (pantomimes condition), 2: a human motion with communicative content (emblems condition) or 3: a non-human motion (scrambled condition)). Participants performed the task with their dominant hand. Accuracy (percentage of correct answers) and response time were measured.

Statistical analysis on behavioral data

Accuracy and response time measures of the biological motion task were analyzed using a 3 x 2 repeated-measures ANOVA with point-light conditions (emblems, pantomimes and scrambled) as within-subjects factor and group (deaf and hearing) as a between-subject factor. A Greenhouse-Geisser correction was applied to the degrees of freedom and to the significance level to prevent the disrespect of the sphericity assumption. Because the duration of the videos varied and ranged from two to four seconds in each point-light condition, a two-way ANOVA was conducted to examine the influence of run (1 and 2) and point-light conditions (emblems, pantomimes, and scrambled stimuli) on stimuli duration. On average, duration time of the videos was 3047.62 ms (SD = 740.013) for emblems, 2857.14 ms (SD = 792.82) for pantomines, and 2380.95 ms (SD = 734.28) for scrambled stimuli. The main effect of the run was not significant (F(2, 120) = .000; p = 1.00), suggesting that the two runs were similar in stimulus duration. However, there was a significant main effect of point-light conditions (F(2, 120) = 10.43; p < .001; η2 = .148) suggesting that the duration of the stimuli differed among to the conditions. The interaction was not significant (F(2, 120) = .000; p > .05). Bonferroni post hoc tests showed that stimulus duration was significantly higher for emblems than for pantomimes (p < .001) and scrambled stimuli (p < .001) whereas no significant differences were found between pantomimes and scrambled stimuli (p > .05). Consequently, these results show that emblem stimuli were significantly longer than the other two conditions. To address this, participants’ response time was transformed into a global mean response time for all point-light conditions across groups. Each response time was then weighted by the duration of the video and multiplied by the global mean.

fMRI acquisition parameters

Whole-brain anatomical and functional images were acquired using a 3-T Trio Tim system (Siemens Magnetom, Erlangen, Germany) equipped with a 32-channel head coil. Multislice T2*-weighted fMRI images were obtained with a gradient echo-planar sequence using axial slice orientation (TR = 2200 ms, TE = 30 ms, FA = 90°, 35 transverse slices, 3.2 mm slice thickness, FoV = 192 x 192 mm2, matrix size = 64 x 64 x 35, voxel size = 3 x 3 x 3.2 mm3). Head movements were restrained using foam pads. A structural T1-weighted MPRAGE sequence was also acquired for all participants (voxel size = 1×1×1 mm3, matrix size = 240 x 256, TR = 2.300 ms, TE = 2.98 ms, TI = 900 ms, FoV 256, 192 slices).

Processing of functional images

The fMRI data were analyzed using SPM 12 in a Matlab environment (Statistical Parametric Mapping, Centre for Neuroimaging, London, UK, http://www.fil.ion.ucl.ac.uk/spm, Matlab 8.5 (Mathworks, Natick, MA, USA). Standard preprocessing was performed (realignment, co-registration of functional and anatomical data). At the step of normalization, two distinct anatomical templates were created using DARTEL [44] (Diffeomorphic Anatomical Registration Through Exponentiated Linear algebra), namely, a template designed for hearing participants and another designed for deaf participants. Both templates were created separately for each group and they have been respectively normalized to the MNI template. A groupwise registration using DARTEL was chosen to reduce possible deformations of the structures that are more difficult to match to the average template based on neurotypical individuals [44]. The DARTEL templates are especially relevant given that previous studies have shown significant structural alterations between deaf and hearing individuals [45]. Finally, spatial smoothing was performed (8-mm FWHM) after which linear contrast images were calculated to test main effects in each participant for each condition ([Emblems], [Pantomimes], [Scrambled]). These linear contrasts generated statistical parametric maps [SPM(T)].

Statistical analyses of fMRI images

The General Linear Model used for the first level analysis (fixed effects) predicted whole brain bold response at each voxel as the dependent variable and conditions: emblem, pantomime and scrambled point-light movements as predictor factors of change in bold response. The resulting individual contrasts, testing the significance of model estimated betas for each condition were smoothed and entered for the second level analysis.

For the second level analysis (random effects), we used the GLM with a full factorial design to estimate the effect of point-light stimulation conditions between groups. The within-subject factor is condition: Emblem, Pantomime, Scrambled. The between-subject factor is group: Deaf vs Hearing. Model estimates resulted in contrasts for the main effect of conditions, the main effect of group and the interactions between groups and conditions. We also tested difference contrasts to assess specific directions of change using t-contrasts. As for biological motion, it was calculated as scrambled contrast subtracted from the sum (Emblems + Pantomimes) in each group separately. We contrasted biological motion, emblems and pantomimes between groups to assess specific activations in the deaf group for the processing of both types of gestures. To assess the relation between change in brain activity and behavioral measures, we used a full factorial design group (2) by condition (3) and difference in response times as a covariate in the model.

Within-group differences

t-contrasts were calculated for the difference between conditions ([Emblems > Pantomimes], [Emblems > Scrambled], [Pantomimes > Emblems], [Pantomimes > Scrambled], [Scrambled > Emblems], [Scrambled > Pantomimes]) using the false discovery rate FDR-correction for multiple comparisons at a probability level p < 0.05. Significantly activated areas are presented as threshold maps in the results section. A conjunction contrast (conjunction null hypothesis) characterized brain areas jointly activated by the contrasts [Emblems + Pantomimes] in both groups.

Between-group analyses

To examine group effect on bold activity change for each condition separately, we calculated t-contrasts for group differences by condition ([Deaf > NH] x [Emblems], [Deaf > NH] x [Pantomimes], ([NH > Deaf] x [Emblems], [NH > Deaf] x [Pantomimes]) using FDR-correction for multiple comparisons, at p < 0.05. The contrast for comparison of brain activations during biological motion processing [(Emblems + Pantomimes)-scrambled] between deaf and NH participants allowed very strict control of low-level stimulus features [19,46].

Finally, we used a general linear model to predict bold change in brain activity with group (deaf vs hearing) and condition (Emblem, Pantomime and Scrambled) as between and within-subjects’ factors respectively. The model included the behavioral differences ([Emblems—Pantomimes]) for response times measure as a covariate. The resulting F-contrast accounted for significant covariance between groups/condition and behavioral differences using FDR-correction for multiple comparisons, at p < 0.05.

Results

Behavioral data

Deaf and hearing groups were equivalent with regards to age (t(30) = .035, p = .682), number of years of education (t(30) = 1.965, p = .06), or on their performance on the fluid reasoning subtest (t(30) = 2.32, p = .43). We performed separate repeated-measures 3 x 2 ANOVAs with both accuracy and response times as the dependent variable. The analysis of correct responses showed a significant main effect of point-light condition (F(1.93, 57.81) = 95.57; p < .001; η2 = .76), no main effect of group (F(1,30) = .04; p = .85) and no significant interaction (F(1.93, 57.81) = 3.08; p > .05). On average, deaf participants recognized 73.38% (SD = 5.33) of emblems correctly, 81.94% (SD = 6.59) of pantomimes and 99.62% (SD = 0.40) of scrambled stimuli as compared to respectively 68.94% (SD = 0.24), 87.69% (SD = 5.11) and 99.56% (SD = 0.65) for hearing participants. Bonferroni post hoc tests demonstrated that all the participants were more accurate in the scrambled condition in comparison to the pantomimes and emblems conditions and more accurate in the pantomimes condition than they were in the emblems condition (p < .001 for all differences).

The analysis of response times showed a significant main effects of point-light condition (F(1.66, 49.88) = 37.69; p < .001; η2 = .56), no significant main effect of group (F(1, 30) = 0.14; p = .71) and, a significant Group × Condition interaction (F(1.66, 49.88) = 4.63; p < .05; η2 = .13) (see Fig 1D). Bonferroni post hoc tests demonstrated that the deaf and hearing participants were fastest at identifying the scrambled condition in comparison to the pantomimes and emblems, respectively (p < .001 for all differences). Only hearing participants exhibited a significant difference between the pantomimes and the emblems conditions, with faster responses for pantomimes (p < .001).

fMRI data

All results reported as significant in this section survived a threshold of whole-brain p < .05 voxel-wise threshold, FWE-corrected. Anatomical labels for active regions are the most probable based on the Harvard-Oxford Cortical Atlas.

Biological versus scrambled motion

We first examined the areas significantly activated by biological motion relative to the scrambled condition [(Emblems + Pantomimes)-scrambled] in each group. As expected, the analyses revealed an overlap in the regions involved in the human action recognition network between the deaf and hearing participants (see Fig 2). Both groups showed extensive bilateral activations that included posterior temporal-occipital regions including V5, pSTS, EBA, and FBA, parietal regions including the right SMG and bilateral SPL; frontal lobe regions including bilateral IFG, frontal operculum/insula, precentral gyrus, middle frontal gyrus, and SMA; and the thalamus bilaterally (see Table 2 for locations of peak activations). Extensive cerebellar activity was observed as well.

Fig 2. fMRI data.

Fig 2

The conjunction of cortical activations implicated in biological motion processing [(Emblems + Pantomimes)—scrambled] by the group, deaf (Red) and hearing participants (Blue), Overlap (Purple).

Table 2. Brain regions showing significant activations for the conjunction of biological motion (emblems and pantomimes)-scrambled in each group.
Anatomical region Hemi Cluster size T x y z p-FEW corr (p < .05) Other areas including Distance (mm)
Deaf
Precentral L 4412 13.63 -54 2 43 .000 Postcentral (4.58)
Frontal mid (10.25)
Fusiform L 5543 12.62 -39 -43 -20 .000 Temporal inf (2.45)
Cerebelum 6 (7.35)
Parietal inf R 65 5.52 30 -46 49 .001 Parietal inf (1.00)
Postcentral (4.58)
Thalamus L 82 5.34 -12 -16 7 .003 Pallidum (11.70)
Caudate (12.04)
Thalamus R 6 4.74 6 -22 -11 .026 Lingual (10.05)
Parahippocampal (10.82)
Fusiform R 3 4.66 36 -4 -41 .034 Temporal inf (2.24)
Temporal mid (6.71)
Hearing
Fusiform R 1766 12.12 39 -43 -20 .000 Temporal inf (5.10)
Cerebelum 6 (5.83)
Temporal mid L 3874 11.80 -51 -70 1 .000 Occipital mid (2.45)
Occipital inf (5.10)
Insula R 1356 10.97 30 26 1 .000 Frontal inf tri (4.58)
Putamen (5.74)
Cerebelum 7b L 473 7.58 -12 -73 -44 .000 Cerebelum 8 (2.24)
Cerebelum crus2 (5.00)
Thalamus L 167 6.14 -9 -16 4 .000 Thalamus R (11.00)
Pallidum (13.45)
Parietal inf R 136 6.10 27 -49 49 .000 Parietal sup (1.73)
Postcentral (5.20)

MNI coordinates (x, y, z) of the significant clusters are given, along with the corresponding brain region for this cluster and the other areas included in each cluster, with distance (mm).

Between-group analyses

Beyond these areas of overlap, some areas showed significant activation only for the deaf group for the biological motion conditions [(Emblems + Pantomimes)-scrambled]. The deaf group showed a significantly stronger bilateral response than the hearing participants in the STG, including the planum temporale (BA 22) and the primary and secondary auditory cortex (BA 41, 42) (see Fig 3B and Table 3). Additionally, only deaf individuals showed activation in the basal ganglia (specifically globus pallidus and the head of the caudate nucleus), and greater extent of activation than hearing individuals in the cerebellum (see Fig 3). In the hearing group, no brain region was found to be more activated than the deaf group (see Table 3).

Fig 3. fMRI data.

Fig 3

(A) The cortical activations implicated in Emblems only (Yellow), Pantomime only (Blue), and the Overlap (Green) by the group. (B) Significant difference between deaf and hearing participants in the biological motion condition, the image in the maximum global coordinate (66.0–28.0 7.0). (C) Significant difference between deaf and hearing participants in the pantomime condition, the image in the maximum global coordinate (66.0–28.0 7.0). (D) Significant difference between deaf and hearing participants in the emblem condition, the image in the maximum peak activation at coordinates (66.0–28.0 7.0). Color scale represents T values (B,C, D).

Table 3. Brain regions showing significant activations for the contrast of Deaf > Hearing in each point-light condition.
Deaf>Hearing
Anatomical region Hemi Cluster size T x y z p-FEW corr (p < .05) Other areas including Distance (mm)
Biological motion
Temporal sup R 969 12.32 66 -28 7 .000 Temporal mid (6.71)
Supramarginal (11.18)
Temporal sup L 916 10.20 -54 -34 10 .000 Temporal mid (2.00)
Supramarginal (8.94)
Precentral L 69 6.46 -57 -1 43 .001 Postcentral (1.73)
Frontal mid (13.96)
Caudate R 26 5.27 18 17 4 .007 Putamen (3.00)
Pallidum (9.00)
Occipital Mid L 32 5.15 -33 -58 7 .005 Calcarine (7.07)
Precuneus (7.35)
Cerebelum 8 L 5 4.75 -3 -61 -32 .005 Vermis 8 (1.41)
Vermis 9 (3.16)
Precentral R 5 4.75 57 8 37 .005 Frontal inf oper (6.16)
Frontal mid (6.78)
Emblems
Temporal sup R 755 10.04 66 -25 4 .000 Temporal Mid (6.08)
Rolandic oper (11.36)
Temporal sup L 819 8.80 -60 -31 7 .000 Supramarginal (9.90)
Precentral L 22 5.50 -54 2 43 .008 Postcentral (4.58)
Frontal Mid (10.25)
Caudate R 2 4.58 18 17 4 .035 Putamen (3.00)
Pallidum (9.00)
Pantomimes
Temporal sup R 213 7.79 66 -28 7 .001 Temporal mid (6.71)
Supramarginal (11.18)
Temporal sup L 144 6.10 -51 -37 10 .002 Temporal mid (2.45)
Rolandic oper (9.49)

MNI coordinates (x, y, z) of the significant cluster are given, along with the corresponding brain region for this cluster and the other areas included in each cluster, with distance (mm).

Brain responses to emblems and pantomimes individually were examined (Deaf > Hearing x [Emblems]; Deaf > Hearing x [Pantomimes]; Hearing > Deaf x [Emblems]; Hearing > Deaf x [Pantomimes]). Again, the deaf group showed a significantly stronger bilateral response to hearing participants in the STG, including the planum temporale (BA 22) and the primary and the secondary auditory cortex (BA 41, 42) (see Fig 4 and Table 3). Notably, the deaf group showed a stronger bilateral response for the emblems condition than for the pantomimes condition, including voxels mostly in the planum temporale and in the primary auditory cortex.

Fig 4. fMRI data.

Fig 4

Covariation between cortical activity triggered by biological motion (Emblems—Pantomimes) and behavioral discrepancy (on reaction times) in the deaf group only. MNI coordinates for global maximum (66.0–28.0 7.0). Graphs: Correlation plots of the blood oxygen level-dependent Emblems-Pantomimes responses in this region against reaction times (RT). Each data point represents a single subject, Red for the deaf group and Green for the hearing group. Color scale represents F values.

Laterality differences in the deaf group

We further investigated whether there were laterality differences within the STG clusters activated uniquely in deaf people. Pairwise comparisons were carried out between the average activity (maximum global coordinate, 66.0–28.0 7.0) in the STG, in both hemispheres in all point-light conditions. The results showed a significant difference in signal strength between the right and the left STG, both in the combined biological motion condition (Emblems + Pantomimes), (t (16) = -8.42, p < .0001 (Right: M ± SD = 2.31 ± 1.07; Left: M ± SD = 1.00 ± .91)) as well as in the emblems condition (t (16) = -5.31, p < .0001 (Right: M ± SD = 2.31 ± .99; Left: M ± SD = 1.22 ± .85)). A rightward asymmetry was found during processing of scrambled motion and emblems but no difference was found between the hemispheres in the pantomime condition (t (16) = -1.41, p > .15 (Right: M ± SD = 2.15 ± 1.24; Left: M ± SD = 1.61 ± .96)). Of interest, an extensive activation of the STG was found in the emblems condition in contrast to the scrambled and pantomimes conditions. The peak activation was located in the primary auditory cortex.

Covariance with behavioral performance

As demonstrated earlier, the behavioral results suggest that there was a significant interaction between hearing status with participants’ response times (Fig 1). Therefore, the way this behavioral difference [Emblems—Pantomimes] translated into neural activations in the deaf group was explored. A factorial model with group (2 levels) and conditions (3 levels) was used in a whole-brain analysis with behavioral differences (Emblems-Pantomimes response times factored out) as covariates. We found a significant covariation between behavioral measures and brain responses in the bilateral STG and the left precentral gyrus (see Fig 4 and Table 4). Brain activity time series in bilateral STG areas were calculated and an independent correlation analysis using Pearson correlation coefficients was carried out to specify the relation between the behavioral measures (response times) and the cerebral activations triggered in the left and right STG. Results indicate that the activation of the STG could predict response times, in the right hemisphere (r = .36, p = .04, R2 = .13) and marginally in the left hemisphere (r = .35, p = .05, R2 = .12). This finding suggests that, for the deaf individuals, stronger activation of the STG during the biological motion task leads to faster response times. Correlation analysis was also conducted on the left precentral gyrus to determine if behavioral results could be predicted by the cortical activity in this region. No significant correlation was found. This was true for the relationship between the peak activity in the precentral gyrus and response times (r = -.37, p > .05) in deaf individuals.

Table 4. Brain regions showing significant activations for the main effect of group with reaction time.
Anatomical region Hemi Cluster size F x y z p-FEW corr (p < .05) Other areas including Distance (mm)
Temporal sup R 108 82.32 66 -28 7 .000 Temporal mid (6.71)
Supramarginal (11.18)
Temporal sup L 140 71.98 -63 -31 7 .000 Supramarginal (9.95)
Precentral L 14 51.40 -57 -1 46 .007 Postcentral (2.45)
Frontal mid (12.88)

MNI coordinates (x, y, z) of the significant cluster are given, along with the corresponding brain region for this cluster and the other areas included in each cluster, with distance (mm).

Discussion

The main goal of the present study was to combine, for the first time, behavioral and neuroimaging measures of emblems and pantomimes gesture recognition, between early-deaf and hearing individuals. In previous studies, inconsistent imaging results were found. A hypoactivation was reported in some cerebral regions involved in the human action network, namely the IPL and the IFG, by two studies investigating the observation of pantomimes in native deaf signers [21,33]. These findings were explained by the predominant use of the visual modality in deaf individuals, not only to support their daily life, but also because of their extensive use of sign language. The latter could be seen as a training in human gestures decoding. The authors argue that this training could make native deaf signers less sensitive to human gestures and thus result in a cortical hypoactivation [33]. More recently, a study looked at congenitally deaf individuals who were native signers [34]. With a pantomime’s judgment task, the authors concluded that there was a robust activation of the human action network in individuals who experienced auditory deprivation in addition to using sign language. However, in this study, no relationship was found between deafs’ linguistic experience and the strength of the cortical activations within the human action recognition network [34]. The present study confirms that there is an overlap in deaf and hearing individuals’ cortical activation network in response to biological motion processing. Both groups showed similar activations in the expected regions [25], that is, occipital, parietal, temporal, and inferior frontal regions during emblems and pantomimes recognition.

More importantly, the present results provide behavioral and neural evidence in favor of compensatory visual cross-modal activity experienced by early deaf people. As some previous studies [14,34,37], we found significant bilateral activations of the STG, including the primary auditory cortex in the deaf group. Our findings corroborate previous work in the literature. Indeed, there are well-established associations between animal and human data [47] showing that deafness can lead to enhanced visual abilities [6,48], thus implying a cross-modal reorganization process where the visual modality recruits the auditory cortex [4,49,50]. However, the evidence is unclear as to whether deafness can lead to both enhanced behavioral performance and a cross-modal activation of the primary auditory cortex by other sensory modalities or higher cognitive functions [1]. Moreover, the literature on the possible behavioral enhancements experienced by deaf individuals is characterized by results that are both heterogeneous and inconsistent. This can be attributed to a variety of factors, such as sample characteristics [48]. Indeed, variables such as the amount of residual hearing, the onset of deafness or etiology of deafness are known to influence the extent of cerebral plasticity [13,51]. Thus, a majority of studies have specifically examined deaf native signers [51], while these deaf individuals represent only a small percentage of the deaf population [52]. Overall, previous results cannot be generalized, and it is therefore complex to have a clear understanding of deaf individuals’ cross-modal reorganization. In our study, differences were found between the behavioral performance and the cortical activation of regions altered by auditory deprivation in deaf compared to hearing participants. The results suggest that early-deaf individuals showed greater sensitivity to the processing of human action than hearing individuals. Specifically, deaf individuals identified emblems as fast as pantomimes in comparison to their hearing peers. These behavioral differences were directly correlated with the bilateral activation of the STG. These results differ from those of previous studies reporting the recruitment of auditory areas in the processing of emblems [37] but not of pantomimes [34], and those reporting no behavioral differences between deaf and hearing participants [34,37]. Additionally, a significant correlation was found between STG activations and response times. This correlation could suggest that the extent of STG recruitment in deaf individuals depends on their capacity to detect emblems more rapidly than pantomimes. This result is consistent with the previous literature showing that enhanced visual performances in deaf individuals are usually related to shorter reaction times rather than to accuracy [5], but must be replicated for exhaustive interpretation.

Furthermore, emblems overall led to more extensive bilateral activations than pantomimes in deaf individuals, especially in the STG (including planum temporale and primary auditory cortex). The activation of the primary auditory cortex, followed by the posterior region of the STG, involved in the dorsal pathway of language processing [5355], suggests that emblems are more prone to be processed as linguistic material by early-deaf individuals. The linguistic processing of emblems, supported by the activation of the left STG, was reported in a study on prelingual deaf adults who were native signers [37]. According to the authors, the linguistic processing of emblems is sustained by a leftward hemispheric asymmetry found in deaf signers in comparison to hearing participants. However, several neuroimaging studies propose that language processing implies a collaboration of both left and right pathways, as well as a cortico-sub-cortical network [53]. In addition, the language network in the right hemisphere is classically related to the visual abilities involved in language processing [56] and explains the STG rightward asymmetry during recognition of visually communicative emblems by the deaf group.

The fMRI analyses performed in the present study addressed the implications of auditory deprivation and linguistic experience on visual biological motion processing. All our deaf participants presented profound-to-severe congenital deafness, but while half of them were proficient in sign language (four were native deaf signers), the other half was using spoken language as a first language. While not formally tested, the robustness of the cortical activations in the human action network suggests an absence of any linguistic experience effect. A particularly interesting finding of the present study is that the differences in human action processing are better explained by an effect of auditory deprivation since all the deaf participants experienced a bilateral activation of the STG. In future studies, a larger sample size of deaf individuals would be needed since deafness related factors are known to influence brain plasticity (e.g. deafness duration, amount of residual hearing, prior use of hearing aids) and should be considered in the analyses [13,51].

Functional and behavioral correlates converge to a human action sensitivity following early-deafness deprivation. This sensitivity does not appear to be modulated by linguistic experience but rather by auditory deprivation. Thus, the present findings are of importance not only because they contribute to the understanding of the visual cross-modal plasticity phenomenon in the deaf population, but also because they offer new avenues of research for rehabilitation strategies that would be better adapted to the daily effects of deafness.

Acknowledgments

The authors declared no competing interests. All data generated or analyzed during this study are included in this published article. We are grateful to the individuals who volunteered for this research and to the staff at the Functional Neuroimaging Unit for testing assistance. We also thank Vanessa Hadid for helpful discussions about data and analysis.

Data Availability

Behavioral Dataset has been shared with a public repository: Simon, Marie (2020): DATASET.xlsx. figshare. https://figshare.com/articles/_/12081405.

Funding Statement

This research was supported by the Canada Research Chair Program (#RGPIN-8245-2014, F.L.), the Canadian Institutes of Health Research (#166197, F.L.), and a grant from Med-EL Elektromed (A.J.N., F.L., and F.C.) and from Quebec Bio-Imaging Network (M.S).” The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

References

  • 1.Merabet LB, Pascual-Leone A. Neural reorganization following sensory loss: the opportunity of change. Nat Rev Neurosci. 2009;11. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2.Alencar CDC, Butler BE, Lomber SG. What and How the Deaf Brain Sees. J Cogn Neurosci [Internet]. 2019. May 21;31(8):1091–109. Available from: 10.1162/jocn_a_01425 [DOI] [PubMed] [Google Scholar]
  • 3.Bavelier D, Tomann A, Hutton C, Mitchell T, Corina D, Liu G, et al. Visual attention to the periphery is enhanced in congenitally deaf individuals. J Neurosci [Internet]. 2000;20(17):RC93 Available from: http://www.ncbi.nlm.nih.gov/pubmed/10952732 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Bavelier D, Hirshorn EA. I see where you’re hearing: how cross-modal plasticity may exploit homologous brain structures. Nat Publ Gr [Internet]. 2010;13(11):1309–11. Available from: 10.1038/nn1110-1309\npapers2:/publication/doi/10.1038/nn1110-1309 [DOI] [PubMed] [Google Scholar]
  • 5.Pavani F, Bottari D. Chapter 22 Visual Abilities in Individuals with Profound Deafness A Critical Review 22.1. Visual abilities in profound deafness: an open challenge for cross-modal plasticity research In: The Neural Bases of Multisensory Processes. CRC Press; 2011. p. 423–48. [Google Scholar]
  • 6.Shiell MM, Champoux F, Zatorre RJ. Enhancement of Visual Motion Detection Thresholds in Early Deaf People. PLoS One [Internet]. 2014;9(2):e90498 Available from: 10.1371/journal.pon [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Heimler B, Weisz N, Collignon O. Revisiting the adaptive and maladaptive effects of crossmodal plasticity. Neuroscience [Internet]. 2014;283:44–63. Available from: 10.1016/j.neuroscience.2014.08.003 [DOI] [PubMed] [Google Scholar]
  • 8.Auer Edward T. J, Bernstein LE. Enhanced Visual Speech Perception in Individuals With Early-Onset Hearing Impairment. J Speech, Lang Hear Res [Internet]. 2007. October 1;50(5):1157–65. Available from: 10.1044/1092-4388(2007/080) [DOI] [PubMed] [Google Scholar]
  • 9.Karns CM, Dow MW, Neville HJ. Altered Cross-Modal Processing in the Primary Auditory Cortex of Congenitally Deaf Adults: A Visual-Somatosensory fMRI Study with a Double-Flash Illusion. J Neurosci [Internet]. 2012. July 11 [cited 2016 Jun 27];32(28):9626–38. Available from: http://www.jneurosci.org/cgi/doi/10.1523/JNEUROSCI.6488-11.2012 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Levänen S, Jousmäki V, Hari R. Vibration-induced auditory-cortex activation in a congenitally deaf adult. Curr Biol [Internet]. 1998. July [cited 2016 Feb 16];8(15):869–72. Available from: http://www.sciencedirect.com/science/article/pii/S096098220700348X [DOI] [PubMed] [Google Scholar]
  • 11.Levänen S, Hamdorf D. Feeling vibrations: Enhanced tactile sensitivity in congenitally deaf humans. Neurosci Lett. 2001;301(1):75–7. 10.1016/s0304-3940(01)01597-x [DOI] [PubMed] [Google Scholar]
  • 12.Fine I, Finney EM, Boynton GM, Dobkins KR. Comparing the effects of auditory deprivation and sign language within the auditory and visual cortex. J Cogn Neurosci [Internet]. 2005;17(10):1621–37. Available from: http://www.mitpressjournals.org/doi/abs/10.1162/089892905774597173%5Cnhttp://www.mitpressjournals.org/doi/pdfplus/10.1162/089892905774597173 [DOI] [PubMed] [Google Scholar]
  • 13.Lambertz N, Gizewski ER, de Greiff A, Forsting M. Cross-modal plasticity in deaf subjects dependent on the extent of hearing loss. Cogn Brain Res [Internet]. 2005. [cited 2017 Apr 6];25(3):884–90. Available from: http://www.sciencedirect.com/science/article/pii/S0926641005002715 [DOI] [PubMed] [Google Scholar]
  • 14.Petitto LA, Zatorre RJ, Gauna K, Nikelski EJ, Dostie D, Evans AC. Speech-like cerebral activity in profoundly deaf people processing signed languages: Implications for the neural basis of human language. Proc Natl Acad Sci [Internet]. 2000. December 5;97(25):13961 LP2–13966. Available from: http://www.pnas.org/content/97/25/13961.abstract [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Sadato N, Okada T, Honda M, Matsuki K-I, Yoshida M, Kashikura K-I, et al. Cross-modal integration and plastic changes revealed by lip movement, random-dot motion and sign languages in the hearing and deaf. Cereb Cortex. 2005. August;15(8):1113–22. 10.1093/cercor/bhh210 [DOI] [PubMed] [Google Scholar]
  • 16.Strelnikov K, Rouger J, Lagleyre S, Fraysse B, Deguine O, Barone P. Improvement in speech-reading ability by auditory training: Evidence from gender differences in normally hearing, deaf and cochlear implanted subjects. Neuropsychologia. 2009. March;47(4):972–9. 10.1016/j.neuropsychologia.2008.10.017 [DOI] [PubMed] [Google Scholar]
  • 17.Cardin V, Smittenaar RC, Orfanidou E, Rönnberg J, Capek CM, Rudner M, et al. Differential activity in Heschl’s gyrus between deaf and hearing individuals is due to auditory deprivation rather than language modality Original Publication: 2016;(124):96–106. [DOI] [PubMed] [Google Scholar]
  • 18.Johansson G. Visual perception of biological motion and a model for its analysis. Percept Psychophys [Internet]. 1973. June;14(2):201–11. Available from: 10.3758/BF03212378 [DOI] [Google Scholar]
  • 19.Zaini H, Fawcett JM, White NC, Newman AJ. Communicative and noncommunicative point-light actions featuring high-resolution representation of the hands and fingers. 2013;319–28. [DOI] [PubMed] [Google Scholar]
  • 20.Brothers L. The neural basis of primate social communication. Motiv Emot. 1990;14(2):81–91. [Google Scholar]
  • 21.Corina D, Chiu Y-S, Knapp H, Greenwald R, San Jose-Robertson L, Braun A. Neural correlates of human action observation in hearing and deaf subjects. Brain Res [Internet]. 2007. [cited 2017 Apr 11];1152:111–29. Available from: http://www.sciencedirect.com/science/article/pii/S0006899307006075 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.Cardin V, Orfanidou E, Kastner L, Ronnberg J, Woll B, Capek CM, et al. Monitoring Different Phonological Parameters of Sign Language Engages the Same Cortical Language Network but Distinctive Perceptual Ones. J Cogn Neurosci. 2016. January;28(1):20–40. 10.1162/jocn_a_00872 [DOI] [PubMed] [Google Scholar]
  • 23.Rizzolatti G, Sinigaglia C. The mirror mechanism: A basic principle of brain function. Nat Rev Neurosci [Internet]. 2016;17(12):757–65. Available from: 10.1038/nrn.2016.135 [DOI] [PubMed] [Google Scholar]
  • 24.Fabbri-Destro M, Rizzolatti G. Mirror Neurons and Mirror Systems in Monkeys and Humans. Physiology [Internet]. 2008. June 1;23(3):171–9. Available from: 10.1152/physiol.00004.2008 [DOI] [PubMed] [Google Scholar]
  • 25.Caspers S, Zilles K, Laird AR, Eickhoff SB. ALE meta-analysis of action observation and imitation in the human brain. Neuroimage [Internet]. 2010. April;50(3):1148–67. Available from: https://linkinghub.elsevier.com/retrieve/pii/S1053811909014013 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26.Saygin AP, Wilson SM, Hagler DJJ, Bates E, Sereno MI. Point-light biological motion perception activates human premotor cortex. J Neurosci. 2004. July;24(27):6181–8. 10.1523/JNEUROSCI.0504-04.2004 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Ozyurek A. Gesture In: Sign language: An international handbook. Mouton; 2012. p. 626–46. [Google Scholar]
  • 28.Arbib MA. From Monkey-like Action Recognition to Human Language: An Evolutionary Framework for Neurolinguistics. Behav Brain Sci. 2004;105–67. [DOI] [PubMed] [Google Scholar]
  • 29.Molnar-Szakacs I, Wu AD, Robles FJ, Iacoboni M. Do you see what I mean? Corticospinal excitability during observation of culture-specific gestures. PLoS One. 2007;2(7):e626 10.1371/journal.pone.0000626 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30.Goldin-Meadow S. The role of gesture in communication and thinking. Trends Cogn Sci. 1999;3(11):419–29. 10.1016/s1364-6613(99)01397-2 [DOI] [PubMed] [Google Scholar]
  • 31.Sandler W, Lillo-martin D. Sign Language and Linguistic Universals Sign Language and Linguistic Universals. 1999;(1991). [Google Scholar]
  • 32.Trettenbrein PC, Papitto G, Zaccarella E. The functional neuroanatomy of sign language in deaf signers: An Activation Likelihood Estimation meta-analysis. 2019;
  • 33.Emmorey K, Xu J, Gannon P, Goldin-meadow S, Braun A. NeuroImage CNS activation and regional connectivity during pantomime observation: No engagement of the mirror neuron system for deaf signers. Neuroimage [Internet]. 2010;49(1):994–1005. Available from: 10.1016/j.neuroimage.2009.08.001 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.Fang Y, Chen Q, Lingnau A, Han Z, Bi Y. Areas Recruited during Action Understanding Are Not Modulated by Auditory or Sign Language Experience. Front Hum Neurosci [Internet]. 2016;10(March):94 Available from: http://www.ncbi.nlm.nih.gov/pubmed/27014025%5Cnhttp://www.pubmedcentral.nih.gov/articlerender.fcgi?artid=PMC4781852 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35.MacSweeney M, Campbell R, Woll B, Giampietro V, David AS, McGuire PK, et al. Dissociating linguistic and nonlinguistic gestural communication in the brain. Neuroimage [Internet]. 2004. [cited 2017 May 17];22(4):1605–18. Available from: http://www.sciencedirect.com/science/article/pii/S1053811904001697 [DOI] [PubMed] [Google Scholar]
  • 36.MacSweeney M, Capek CM, Campbell R, Woll B. The signing brain: the neurobiology of sign language. Vol. 12, Trends in Cognitive Sciences. 2008. p. 432–40. [DOI] [PubMed] [Google Scholar]
  • 37.Husain FT, Patkin DJ, Kim J, Braun AR, Horwitz B. Dissociating neural correlates of meaningful emblems from meaningless gestures in deaf signers and hearing non-signers. Brain Res. 2012; [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 38.Husain FT, Patkin DJ, Thai-Van H, Braun AR, Horwitz B. Distinguishing the processing of gestures from signs in deaf individuals: An fMRI study. Brain Res [Internet]. 2009;1276:140–50. Available from: http://www.sciencedirect.com/science/article/pii/S0006899309007859 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 39.Capek CM, Woll B, MacSweeney M, Waters D, McGuire PK, David AS, et al. Superior temporal activation as a function of linguistic knowledge: insights from deaf native signers who speechread. Brain Lang. 2010. February;112(2):129–34. 10.1016/j.bandl.2009.10.004 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 40.Oldfield RC. The assessment and analysis of handedness: The Edinburgh inventory. Neuropsychologia [Internet]. 1971. [cited 2017 Jun 20];9(1):97–113. Available from: http://www.sciencedirect.com/science/article/pii/0028393271900674 [DOI] [PubMed] [Google Scholar]
  • 41.Wechsler D, Hsiao-pin C. NCS Pearson; San Antonio, TX: 2011. WASI-II Wechsler Abbreviated Scale Intell.
  • 42.Strauss E, Sherman EMS, Spreen O. A compendium of neuropsychological tests: Administration, norms, and commentary. American C. 2006.
  • 43.Newman AJ, Supalla T, Fernandez N, Newport E, Bevelier D. Neural systems supporting linguistic structure, linguistic experience, and symbolic communication in sign language and gesture. Proc Natl Acad Sci. 2015;112(37):11684–9. 10.1073/pnas.1510527112 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 44.Ashburner J. A fast diffeomorphic image registration algorithm. 2007;38:95–113. [DOI] [PubMed] [Google Scholar]
  • 45.Simon M, Campbell E, Genest F, MacLean MW, Champoux F, Lepore F. The Impact of Early Deafness on Brain Plasticity: A Systematic Review of the White and Gray Matter Changes [Internet]. Vol. 14, Frontiers in Neuroscience. 2020. p. 206 Available from: https://www.frontiersin.org/article/10.3389/fnins.2020.00206 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 46.Peuskens H, Vanrie J, Verfaillie K, Orban GA. Specificity of regions processing biological motion. Eur J Neurosci [Internet]. 2005. May 27;21(10):2864–75. Available from: 10.1111/j.1460-9568.2005.04106.x [DOI] [PubMed] [Google Scholar]
  • 47.Lomber SSG, Meredith MA, Kral A. Cross-modal plasticity in specific auditory cortices underlies visual compensations in the deaf. Nat Neurosci [Internet]. 2010;13:1421–7. Available from: http://www.ncbi.nlm.nih.gov/pubmed/20935644%5Cnhttp://www.nature.com/neuro/journal/v13/n11/abs/nn.2653.html [DOI] [PubMed] [Google Scholar]
  • 48.Pavani F, Bottari D. Visual Abilities in Individuals with Profound Deafness. In: The Neural Bases of Multisensory Processes [Internet]. CRC Press/Taylor & Francis; 2011. [cited 2016 Sep 28]. p. 423–48. http://www.ncbi.nlm.nih.gov/pubmed/22593887 [Google Scholar]
  • 49.Bottari D, Heimler B, Caclin A, Dalmolin A, Giard M-H, Pavani F. Visual change detection recruits auditory cortices in early deafness. Neuroimage. 2014;94:172–84. 10.1016/j.neuroimage.2014.02.031 [DOI] [PubMed] [Google Scholar]
  • 50.Shiell MM, Champoux F, Zatorre RJ. The Right Hemisphere Planum Temporale Supports Enhanced Visual Motion Detection Ability in Deaf People: Evidence from Cortical Thickness. Neural Plast [Internet]. 2016. January 14;2016:7217630. Available from: http://www.ncbi.nlm.nih.gov/pmc/articles/PMC4738967/ [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 51.Bavelier D, Dye MWG, Hauser PC. Do deaf individuals see better? Trends Cogn Sci. 2006;10(11):512–8. 10.1016/j.tics.2006.09.006 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 52.Mitchell RE, Young T a., Bachleda B, Karchmer M a. How Many People Use ASL in the United States? Why Estimates Need Updating. Sign Lang Stud. 2006;6(3):306–35. [Google Scholar]
  • 53.Dick AS, Bernal B, Tremblay P. The language connectome: new pathways, new concepts. Neuroscientist. 2014. October;20(5):453–67. 10.1177/1073858413513502 [DOI] [PubMed] [Google Scholar]
  • 54.Friederici AD. The brain basis of language processing: from structure to function. Physiol Rev. 2011. October;91(4):1357–92. 10.1152/physrev.00006.2011 [DOI] [PubMed] [Google Scholar]
  • 55.Friederici AD, Gierhan SME. The language network. Curr Opin Neurobiol. 2013. April;23(2):250–4. 10.1016/j.conb.2012.10.002 [DOI] [PubMed] [Google Scholar]
  • 56.Purves D. Neuroscience. Sunderland, Mass: Sinauer Associates; 2012. [Google Scholar]

Decision Letter 0

Paul Hinckley Delano

4 Feb 2020

PONE-D-19-34925

Enhancement of Visual Biological Motion Recognition in Early-Deaf Adults: Functional and Behavioral Correlates

PLOS ONE

Dear Mrs Simon,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

We would appreciate receiving your revised manuscript by Mar 20 2020 11:59PM. When you are ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter.

To enhance the reproducibility of your results, we recommend that if applicable you deposit your laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). This letter should be uploaded as separate file and labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. This file should be uploaded as separate file and labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. This file should be uploaded as separate file and labeled 'Manuscript'.

Please note while forming your response, if your article is accepted, you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out.

We look forward to receiving your revised manuscript.

Kind regards,

Paul Hinckley Delano, Ph.D.

Academic Editor

PLOS ONE

Journal Requirements:

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at

http://www.journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and http://www.journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. Thank you for stating the following financial disclosure:

"This research was supported by the Canada Research Chair Program (F.L.), the Natural Sciences and Engineering Research Council of Canada (F.L.), and a grant from Med-EL Elektromed (A.J.N., F.L., and F.C.) and from Quebec Bio-Imaging Network (M.S). "           

Please state what role the funders took in the study.  If the funders had no role, please state: "The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript."

If this statement is not correct you must amend it as needed.

Please include this amended Role of Funder statement in your cover letter; we will change the online submission form on your behalf.

3. Please provide further details regarding how participants were recruited, including the participant recruitment date.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Yes

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: Yes

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: The work contrast the recognition of different inferred movements on deaf and hearing population. The study aims to study is rooted not only in exploring cortical processing related to these movements and population characteristics, but also to link this cortical processing with behavioral differences.

In general, the article concludes that minor behavioral differences are found, which are weakly related to STG activity. Cortical activity presents evident differences between groups and conditions.

This is a well written article with a robust experiment design. Methodologically speaking it presents a classic approach properly applied.

However, there are still some room for improvement. The article may improve a few things to facilitate reader’s comprehension, and in my personal opinion, this article may use some different methodological approaches to improve the results they based on in part of the discussion. I think that authors have better results than what they think, but the technique used, does not allow them to catch such results entirely. Finally, there are some methodological decisions that are not argued and they should be. Main reasons, is to improve the understanding of the logic behind those calls, but also to improves readers’ experience and allow people who are naïve to statistics to understand the justifications of the method used.

Specifically:

1. Some results are given in methods sections, which I don’t clearly see why aren’t they in results section.

2. I highly suggest to change “condition” for a meaningful term such a movements or anything that suits authors idea of what they are measuring through those three conditions. This will improve readers’ experience.

3. In results given within methods section, mean +- SD was not reported. Please complete the statistical reports.

4. Authors indicates that they used “global mean response”. I did not understand properly this normalization with the information given in the text. I search information about it, and I find many different conceptualizations for the same concept. As such, I would highly appreciate to include a detailed explanation of what was made in order to increase replicability of this work. Particularly the weighting method is not mentioned at all. Also, no cite for this method, or rationale behind it is given. As it is currently given, it is not easy to understand the properties of this normalization or the reasons to use it.

5. Also, it is not clear when this normalization is used, as in Figure 1 Reaction times are ploted raw.

6. It is not justified why using the differences in reaction time (unclear if normalized or not) as covariates. I also did not understand in which statistical model they were introduce as covariates (which was the dependent variable). Since it is mentioned as start of new paragraph, I don’t understand if it refers to previous or following analyses. I assume, it refers to repeated measures ANOVA; however, this cannot be assessed with ANCOVA, as there is not properly developed repeated measures ANCOVA. This is usually assessed using Mixed Linear Models. In any case, there is no justification to include it in any of them as covariate. I would appreciate a clarification of this point.

7. In line 244 authors say that they will use a correlation (without describing which method specifically) using an F-test. From results I induce that the authors used Pearson correlation, which used r-statistic instead of F statistic. F-statistic is used to evaluate the significance of a linear model, using t-statistic for each coefficient. Please clarify what are the authors exactly applying.

8. In line 261-2 authors cite Figure 1E, but it only has until D.

9. Figure 1 Caption appears 2 times in text body.

10. In Table 2, authors wrote “most significant”. This is vague. Please explain the criteria by which those results were included and why some others were excluded of the report.

11. In line 350, authors wrote “STG significantly predicted response time” base on (what I induce) a correlation. This is not appropriate. Correlations do not allow to talk about predictions or estimations, as coefficients are not estimated. Strictly speaking in this study only estimations can be made as there is no train and new data logic. I don’t mind the misusage of prediction as it may contribute to improves readers’ understanding of the work (most people in biological areas misused the concept), but then coefficients must be reported.

12. In Figure 4 please include a color legend to improve reader’s experience.

13. Authors should consider the usage of F-test in correlations presented in Figure 4, considering that here authors imply a hypothesis that neural activity produce behavior (which makes total sense). In that case, it is better suited a linear model. Also, and considering the plots, particularly for left hemisphere, I would advise authors to fit a linear model and to remove outliers base on bonferroni’s outlier test (is model based, so authors must fit first the model, remove outliers and fit again). Author may also try to fit a model in y=b0+b(Reaction Time)+b(Reaction Time)squared. This would allow to capture the parabolic like pattern that this curve has in left hemisphere in deaf group. It is likely that increase in activity plateaus at some point, which would be consistent with non-linear modeling. This similar pattern seem to occur in NH for right STG. I would also encourage to include other relevant variables in the model which may contribute to reduce the noise. Personally, I think that the method used is not taking the best of the results obtained.

14. Finally, squared Rs derived from correlations are too weak to establish conclusive results. As such, depending on how linear modeling goes, authors may use bootstrap or any cross validation technique to evaluate if fit is low but consistent. Otherwise, I would expect to moderate discussion considering that this result may not be obtained if the study is replicated.

Reviewer #2: The authors study behavioral and fMRI measures of emblems and pantomines gestures recognition, between early-deaf and hearing individuals.

The total number of participants is 32 (16 early-deaf and 16 hearing subjects matched on age, sex and number of education years).

The article is very clear and well written. The methodology is sound, with stimuli that were carefully prepared and the results were appropiately analyzed. Results are well presented and very interesting.

A main finding is the greater sensitivity to the processing of human action in early-deaf invididuals (with a processing of emblems as fast as patomines) that is better explained by an effect of auditory deprivation. These behavioral differences where directly correlated with a bilateral activation of the STG. Also a correlation was found between these activations and response time, suggesting that the extent of STG recruitment in deaf subjects depends on the capacity that have to detect emblems more rapidly than pantomine.

My only concern is the size of the sample. If the paper is accepted, the authors should discuss the possible effects of using such a small sample... for example, there were no differences for the primary language of the deaf individuals (sign, native sign and spoken)... could this be due to the use of too small a sample?

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files to be viewed.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email us at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2020 Aug 10;15(8):e0236800. doi: 10.1371/journal.pone.0236800.r002

Author response to Decision Letter 0


3 Apr 2020

We are happy to enclose a revised version of the manuscript PONE-D-19-34925 entitled “Enhancement of Visual Biological Motion Recognition in Early-Deaf Adults: Functional and Behavioral Correlates” for consideration in Plos One.

We have carefully taken all the reviewer's suggestions into consideration to prepare a revised version of the manuscript, which also required modifications in the text and a revised figure. We truly feel the reviewers have contributed to improving the quality and clarity of the manuscript. Therefore, we would like to thank both reviewers for their precise and useful comments and suggestions as well as for their time spent reviewing our work.

Below, you can find our responses to the comments of Reviewers 1 and 2 as well as actions taken in the the revised manuscript.

We hope that you will find our revised manuscript satisfactory for publication and would be very proud to be given the opportunity to represent your journal.

Report of Reviewer #1

The work contrasts the recognition of different inferred movements on deaf and hearing population. The study aims to study is rooted not only in exploring cortical processing related to these movements and population characteristics, but also to link this cortical processing with behavioral differences.

In general, the article concludes that minor behavioral differences are found, which are weakly related to STG activity. Cortical activity presents evident differences between groups and conditions.

This is a well written article with a robust experiment design. Methodologically speaking it presents a classic approach properly applied.

However, there are still some room for improvement. The article may improve a few things to facilitate reader’s comprehension, and in my personal opinion, this article may use some different methodological approaches to improve the results they based on in part of the discussion. I think that authors have better results than what they think, but the technique used, does not allow them to catch such results entirely. Finally, there are some methodological decisions that are not argued, and they should be. Main reasons are to improve the understanding of the logic behind those calls, but also to improves readers’ experience and allow people who are naïve to statistics to understand the justifications of the method used.

Specifically:

1. Some results are given in methods sections, which I don’t clearly see why aren’t they in results section.

Response: Thank you for your pertinent comment. As you mentioned, some results were given in the method section. Thus, we removed the following sentence: “Deaf and hearing groups were equivalent with regards to age (t(30) = .035, p = .682), number of years of education (t(30) = 1.965, p = .06), or on their performance on the fluid reasoning subtest (t(30) = 2.32, p = .43)”, previously L.161 to L255, in the beginning of the method.

2. I highly suggest to change “condition” for a meaningful term such a movements or anything that suits authors idea of what they are measuring through those three conditions. This will improve readers’ experience.

Response: We thank the reviewer for the suggestion and have adjusted the term “condition” by “point-light condition” in all the manuscript. The new terminology has been inspired by, previous studies such as. Campbell et al., 2011.

3. In results given within methods section, mean +- SD was not reported. Please complete the statistical reports.

Response: Thank you for your observation. We added all the mean+/- SD missing in the first draft of the manuscript.

4. Authors indicates that they used “global mean response”. I did not understand properly this normalization with the information given in the text. I search information about it, and I find many different conceptualizations for the same concept. As such, I would highly appreciate to include a detailed explanation of what was made in order to increase replicability of this work. Particularly the weighting method is not mentioned at all. Also, no cite for this method, or rationale behind it is given. As it is currently given, it is not easy to understand the properties of this normalization or the reasons to use it.

Response: As explained in the manuscript (starting L.198), we used normalized reaction times because 1) we found a significant difference between the time duration of the three types of stimuli, 2) despite the instruction to press as fast and accurately as possible, all the participant (deaf and hearing) waiting for the end of the stimulus to answer. Therefore, we normalized reaction times based on a calculation below: each individual's response time to a video has been divided by the duration of that specific video then multiplied by the global mean, which represents the mean duration of all 126 videos combined (L212).

5. Also, it is not clear when this normalization is used, as in Figure 1 Reaction times are plotted raw.

Response: As explained above, we chose to normalize reaction times because the duration of the videos varied from two to four seconds in each point-light condition. Thus, Figure 1 represents normalized reaction times and all following statistical tests on the reaction time measure have been performed with these. We are confident that they represent an optimal measure of behavior during the recognition of biological motion.

6. It is not justified why using the differences in reaction time (unclear if normalized or not) as covariates. I also did not understand in which statistical model they were introduce as covariates (which was the dependent variable). Since it is mentioned as start of new paragraph, I don’t understand if it refers to previous or following analyses. I assume, it refers to repeated measures ANOVA; however, this cannot be assessed with ANCOVA, as there is not properly developed repeated measures ANCOVA. This is usually assessed using Mixed Linear Models. In any case, there is no justification to include it in any of them as covariate. I would appreciate a clarification of this point.

Response: The difference in reaction times for emblems and pantomimes is not normalized because reaction times were already normalized to total mean response time for the three conditions and video duration. There is also a significant difference in response times for both groups between conditions (behavioral differences statistical analysis). This difference accounts for the additional semantic processing time of emblems compared to pantomimes. The difference in RTs is then introduced in the 2nd level (fMRI statistical analysis in SPM) regression model as a covariate to test correlation between reaction time change and brain signal change in Emblem-pantomime contrast across subjects.

7. In line 244 authors say that they will use a correlation (without describing which method specifically) using an F-test. From results I induce that the authors used Pearson correlation, which used r-statistic instead of F statistic. F-statistic is used to evaluate the significance of a linear model, using t-statistic for each coefficient. Please clarify what are the authors exactly applying.

Response: In this case, the 2nd level analysis in SPM was applied using a multiple regression model introducing signal change for emblem-pantomime contrast as the dependent variable and the difference in reaction time between emblem and pantomime as a covariate for both groups, to test for covariation (either positive or negative) between behavioral performance and brain signal change.

The t statistic is used to test for Emblem-Pantomime contrast and the covariance with RT difference. As such for voxels which pass correction the signal change and performance are significantly correlated. The results in Figure 4 graph represent the relationship between behavioral change and brain activity at the voxel of maximum activity which value was extracted from SPM.

8. In line 261-2 authors cite Figure 1E, but it only has until D.

Response: We thank the reviewer for the suggestion and have corrected this error and replaced “Figure 1E” by “Figure 1D”.

9. Figure 1 Caption appears 2 times in text body.

Response: Thank you for your comment. We remove one on the legend under the name of “Figure 1” in the main text.

10. In Table 2, authors wrote “most significant”. This is vague. Please explain the criteria by which those results were included and why some others were excluded of the report.

Response: Thank you for your comment. Indeed, we re-phrased this sentence, instead of “MNI coordinates (x, y, z) of the most significant cluster are given (…)”,

We suggest: “MNI coordinates (x, y, z) of the significant cluster are given (…).” Thus, we applied this correction to all the MRI legends.

11. In line 350, authors wrote “STG significantly predicted response time” base on (what I induce) a correlation. This is not appropriate. Correlations do not allow to talk about predictions or estimations, as coefficients are not estimated. Strictly speaking in this study only estimations can be made as there is no train and new data logic. I don’t mind the misusage of prediction as it may contribute to improves readers’ understanding of the work (most people in biological areas misused the concept), but then coefficients must be reported.

Response: We thank the reviewer for the suggestion. We have strictly pursued statistical guidelines for correlations based on Cohen, J. 1988. Statistical Power Analysis for the Behavioral Sciences, 2nd Edition. Routledge. Thus, L. 372, we have reported all classical values: p, r and R2 as coefficient. We re-phrased this sentence, instead of “STG significantly predicted response time”.

We suggest: “STG could predict response time”.

12. In Figure 4 please include a color legend to improve reader’s experience.

Response: Thank you, we agree with your suggestion. We have proceeded to minor changes in the Fig.4 to include a color legend. We believe this has helped to clarify the figure.

13. Authors should consider the usage of F-test in correlations presented in Figure 4, considering that here authors imply a hypothesis that neural activity produce behavior (which makes total sense). In that case, it is better suited a linear model. Also, and considering the plots, particularly for left hemisphere, I would advise authors to fit a linear model and to remove outliers base on bonferroni’s outlier test (is model based, so authors must fit first the model, remove outliers and fit again). Author may also try to fit a model in y=b0+b (Reaction Time)+b(Reaction Time)squared. This would allow to capture the parabolic like pattern that this curve has in left hemisphere in deaf group. It is likely that increase in activity plateaus at some point, which would be consistent with non-linear modeling. This similar pattern seems to occur in NH for right STG. I would also encourage to include other relevant variables in the model which may contribute to reduce the noise. Personally, I think that the method used is not taking the best of the results obtained.

Response: As explained above the statistical analysis for covariance between RT difference and brain activity was done in SPM as part of the imaging data statistical analysis. The use of a linear model is interesting but as long as there is a collinearity between brain activation and behavioral performance other variables are introduced instead of STG brain activation. The communication means (sign, aural) seems to predict well the change in RT. With signers showing reduced change in RTs between Emblems and pantomimes. See results:

14. Finally, squared Rs derived from correlations are too weak to establish conclusive results. As such, depending on how linear modeling goes, authors may use bootstrap or any cross-validation technique to evaluate if fit is low but consistent. Otherwise, I would expect to moderate discussion considering that this result may not be obtained if the study is replicated.

Response: We will consider discussing this point in the paper, see L. 440-447.

Report of Reviewer #2

The authors study behavioral and fMRI measures of emblems and pantomimes gestures recognition, between early-deaf and hearing individuals. The total number of participants is 32 (16 early-deaf and 16 hearing subjects matched on age, sex and number of education years).

The article is very clear and well written. The methodology is sound, with stimuli that were carefully prepared, and the results were appropriately analyzed. Results are well presented and very interesting.

A main finding is the greater sensitivity to the processing of human action in early-deaf individuals (with a processing of emblems as fast as pantomimes) that is better explained by an effect of auditory deprivation. These behavioral differences where directly correlated with a bilateral activation of the STG. Also, a correlation was found between these activations and response time, suggesting that the extent of STG recruitment in deaf subjects depends on the capacity that have to detect emblems more rapidly than pantomime.

My only concern is the size of the sample. If the paper is accepted, the authors should discuss the possible effects of using such a small sample... for example, there were no differences for the primary language of the deaf individuals (sign, native sign and spoken)... could this be due to the use of too small a sample?

Response: Thank you for your comment, we agree that this information should be mentioned as one of the conclusions and have modified the manuscript to account for your comment. To express this, we suggest adjusting the wording with the following: “In future studies, a larger sample size of deaf individuals would be needed since deafness related factors influencing brain plasticity (e.g. deafness duration, amount of residual hearing, prior use of hearing aids) and should be considered in the analyses (13,54).” However, we are very confident with our result showing that behavioral difference was correlated to a bilateral activation in the STG in all our early deaf participants, independently of the primary mean of communication.

Finally, we want to report to the reviewer 2 the challenges that are linked to the recruitment of deaf individuals, especially for imaging research who are, at some point, disconnected to their daily-life preoccupation. Therefore, the specific characteristics of this population make it difficult to access through the usual recruitment “channels” despite support, in our case, from a major deaf rehabilitation center in Montreal. In previous studies, the number of deaf adults tested in imaging varied between 6 to 53 (see for example, Meyer et al., 2007 or Shibata 2007, Lepore et al., 2010 or Smithenaar et al., 2016). We recently performed a power analysis to estimate the optimal sample. The analysis suggests that a sample of 25 deaf participants is required to perform linear regression linked to deafness heterogeneity, thus, we hope that futures studies will replicate our main results while enriching them with regression analyzes.

Attachment

Submitted filename: Response to Reviewers.docx

Decision Letter 1

Paul Hinckley Delano

1 May 2020

PONE-D-19-34925R1

Enhancement of Visual Biological Motion Recognition in Early-Deaf Adults: Functional and Behavioral Correlates

PLOS ONE

Dear Mrs Simon,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please respond to all concerns raised by reviewer 1.

We would appreciate receiving your revised manuscript by Jun 15 2020 11:59PM. When you are ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter.

To enhance the reproducibility of your results, we recommend that if applicable you deposit your laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). This letter should be uploaded as separate file and labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. This file should be uploaded as separate file and labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. This file should be uploaded as separate file and labeled 'Manuscript'.

Please note while forming your response, if your article is accepted, you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out.

We look forward to receiving your revised manuscript.

Kind regards,

Paul Hinckley Delano, Ph.D.

Academic Editor

PLOS ONE

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #1: All comments have been addressed

Reviewer #2: All comments have been addressed

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: Authors have addressed most of the comments given. However, there are still one major confusion which limits strongly the replicability of the reported results, and two minor comments about the Figures.

1- In line 251 to 254, the GLM method is briefly explained. First, there is no mention to the specific GLM used. From comments of the authors, I must assume that a Generalized Mixed Linear Model was used. Second, authors mention correlations (without mention which kind) analysis as synonym of GLM, which is not the case. Third, there is no mention to the structure of the GLMM used. Which variables were of first level, second level, were interaction specified? which ones, and so on. This is relevant, as introducing random effects and cross-level interactions are likely to produce overfit. Fourth, authors report Pearson r instead of beta coefficient, while saying that they used GLM. GLM uses F-statistic, not r-statistic. One can obtain r.-statistic from GLM; however, is not a proper results report and omit the most relevant information, like the beta value.

Since authors have been so consistent in using as synonym correlations, GLM and GLMM, I reviewed in detail the SPM documentation and documentation used for tutorial and divulgation purposes. In many of these documents it is stated that GLM implemented in SPM can be used to asses t-test, correlations, linear modeling, and generalized linear models. Correlations and t-test are not GLM; however, one can used particular cases of GLM to answer almost the same questions answered by t-test and correlations. In the case of ANOVA and ANCOVA, they are fairly the same test. However, in the case of correlation and t-test they are not equivalent to GLM. Discrepancies will respond mostly to assumption violation and sample size used. Suggesting them as synonym only contributes to obscure the methods used and limiting replicability.

I understand the source of the confusion. However, current statistical report limits strongly replicability and metanalytic work. To solve the current confusion, please:

a-Refer always to GLM always instead of correlation.

b-In method sections specify properly the structure of the GLM model used. Which variables in 1st and 2nd level, if any interaction was introduced, and which variables were introduced as fixed or random effects.

c-Please check the SPM output if beta values (also referred as coefficients) are reported. It may be the case that if you are using random effects, these will not be reported, or they would not be of easy access. In that case, please remove the r-statistics of the current results report (p-values are not estimated based on such r-statistic, therefor is misleading to report it).

d-If possible, report the F-statistic from which GLM model significance is estimated (current draft does not report GLM significance, I recommend including it if possible). Also, if possible, include the t statistic usually used for GLMM coefficients.

2- In Figure 3 B, C, D and in Figure 4, there is a yellow to white color scale without units. There is no explanation either of what these scales represents. Please include the vriable to which is referring and the units of such variables.

3- Figure 1 D, standard unit for milliseconds is ms, not MSEC. Capital M states Mega, which means *10^6, while m means *10^-3. Sec is an informal abbreviation; s is the standard unit for seconds.

Reviewer #2: (No Response)

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files to be viewed.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email us at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2020 Aug 10;15(8):e0236800. doi: 10.1371/journal.pone.0236800.r004

Author response to Decision Letter 1


14 Jun 2020

We are happy to enclose a second revised version of the manuscript PONE-D-19-34925 entitled “Enhancement of Visual Biological Motion Recognition in Early-Deaf Adults: Functional and Behavioral Correlates” for consideration in Plos One.

We have carefully taken all the reviewer 1 suggestions into consideration to prepare a revised version of the manuscript, which also required modifications in the text and a revised figure. We truly feel the reviewer have contributed to improving the quality and clarity of the manuscript.

Below, you can find our responses to the comments of Reviewer 1 as well as actions taken in the revised manuscript.

We hope that you will find our revised manuscript satisfactory for publication and would be very proud to be given the opportunity to represent your journal.

Report of Reviewer #1

Reviewer #1: Authors have addressed most of the comments given. However, there are still one major confusion which limits strongly the replicability of the reported results, and two minor comments about the Figures.

1- In line 251 to 254, the GLM method is briefly explained. First, there is no mention to the specific GLM used. From comments of the authors, I must assume that a Generalized Mixed Linear Model was used. Second, authors mention correlations (without mention which kind) analysis as synonym of GLM, which is not the case. Third, there is no mention to the structure of the GLMM used. Which variables were of first level, second level, were interaction specified? which ones, and so on. This is relevant, as introducing random effects and cross-level interactions are likely to produce overfit. Fourth, authors report Pearson r instead of beta coefficient, while saying that they used GLM. GLM uses F-statistic, not r-statistic. One can obtain r.-statistic from GLM; however, is not a proper results report and omit the most relevant information, like the beta value.

Since authors have been so consistent in using as synonym correlations, GLM and GLMM, I reviewed in detail the SPM documentation and documentation used for tutorial and divulgation purposes. In many of these documents it is stated that GLM implemented in SPM can be used to asses t-test, correlations, linear modeling, and generalized linear models. Correlations and t-test are not GLM; however, one can used particular cases of GLM to answer almost the same questions answered by t-test and correlations. In the case of ANOVA and ANCOVA, they are fairly the same test. However, in the case of correlation and t-test they are not equivalent to GLM. Discrepancies will respond mostly to assumption violation and sample size used. Suggesting them as synonym only contributes to obscure the methods used and limiting replicability.

I understand the source of the confusion. However, current statistical report limits strongly replicability and metanalytic work. To solve the current confusion, please:

a-Refer always to GLM always instead of correlation.

Response: Thank you for your pertinent comment, this is now done when it applies.

b-In method sections specify properly the structure of the GLM model used. Which variables in 1st and 2nd level, if any interaction was introduced, and which variables were introduced as fixed or random effects.

Response: Thank you, we agree with your suggestion. We have proceeded to minor changes This section is now part of the manuscript:

The General Linear Model for first level analysis predicted the bold response as the dependent variable and conditions: emblem, pantomime and scrambled point light movements as predictor factors of change in bold response.

The resulting individual contrasts (comparing estimated Betas to 0 using one tailed t-test) for each condition were smoothed and entered for the second level analysis.

For the second level analysis we used a full factorial design to estimate the effect of point light stimulation conditions between groups. The within subject factor is condition: Emblem, Pantomime, Scrambled. The between subject factor is group: Deaf vs controls.

Model estimates resulted in contrasts for the main effect of conditions, the main effect of group and the interactions between groups and conditions.

We also tested difference contrasts to assess specific directions of change using t-tests. Biological motion contrast was calculated as scrambled subtracted from the sum (Emblems + Pantomimes) in each group separately.

We contrasted biological motion, emblems and pantomimes between groups to assess specific activations in the deaf group for the processing of both types of gestures.

To assess the relation between change in brain activity and behavioral measures, we used a full factorial design group (2) by condition (3) and difference in response times as a covariate in the model.

c-Please check the SPM output if beta values (also referred as coefficients) are reported. It may be the case that if you are using random effects, these will not be reported, or they would not be of easy access. In that case, please remove the r-statistics of the current results report (p-values are not estimated based on such r-statistic, therefor is misleading to report it).

Response: Once the model of covariation between changes in brain activity between groups and conditions and behavioral measures was estimated, areas showing significant covaration with performance measures were identified in bilateral STG regions.

Resulting brain activity time series in bilateral STG areas were calculated (at peak activity) and an independent correlation analysis was carried out using Pearson correlation coefficients to specify the relation between the behavioral measures (response times) and the cerebral activations in the left and right STG. Correlation results reported in the manuscript are related to this analysis.

This is now precised in the manuscript by adding a section.

d-If possible, report the F-statistic from which GLM model significance is estimated (current draft does not report GLM significance, I recommend including it if possible). Also, if possible, include the t statistic usually used for GLMM coefficients.

Response:

F statistic for the main effect of group: F=22.5960 (p=0.05 FDR corrected)

F statistic for the main effect of condition: F=13.6698 (p=0.05 FDR corrected)

F statistic for the covariance with behavioral performance F=35.3696 (p=0.05 FDR corrected)

Minor

2- In Figure 3 B, C, D and in Figure 4, there is a yellow to white color scale without units. There is no explanation either of what these scales represent. Please include the variable to which is referring and the units of such variables.

Response: Thank you for your observation, we added the missing information.

Figure 3: color scale represents T values

Figure 4: color scale represents F values

3- Figure 1 D, standard unit for milliseconds is ms, not MSEC. Capital M states Mega, which means *10^6, while m means *10^-3. Sec is an informal abbreviation; s is the standard unit for seconds.

Response: We thank the reviewer for the suggestion and have corrected this error in the Figure 1D.

Attachment

Submitted filename: Response to Reviewers.docx

Decision Letter 2

Paul Hinckley Delano

16 Jun 2020

PONE-D-19-34925R2

Enhancement of Visual Biological Motion Recognition in Early-Deaf Adults: Functional and Behavioral Correlates

PLOS ONE

Dear Dr. Simon,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

As suggested by one reviewer, give a rigorous description of all statistical tests and fix minor issues in Figures.

Please submit your revised manuscript by Jul 31 2020 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols

We look forward to receiving your revised manuscript.

Kind regards,

Paul Hinckley Delano, Ph.D.

Academic Editor

PLOS ONE

[Note: HTML markup is below. Please do not edit.]

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

Decision Letter 3

Paul Hinckley Delano

15 Jul 2020

Enhancement of Visual Biological Motion Recognition in Early-Deaf Adults: Functional and Behavioral Correlates

PONE-D-19-34925R3

Dear Dr. Simon,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Paul Hinckley Delano, Ph.D.

Academic Editor

PLOS ONE

Additional Editor Comments (optional):

Reviewers' comments:

Acceptance letter

Paul Hinckley Delano

20 Jul 2020

PONE-D-19-34925R3

Enhancement of Visual Biological Motion Recognition in Early-Deaf Adults: Functional and Behavioral Correlates

Dear Dr. Simon:

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department.

If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org.

If we can help with anything else, please email us at plosone@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Paul Hinckley Delano

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    Attachment

    Submitted filename: Response to Reviewers.docx

    Attachment

    Submitted filename: Response to Reviewers.docx

    Attachment

    Submitted filename: Response to Reviewers.docx

    Data Availability Statement

    Behavioral Dataset has been shared with a public repository: Simon, Marie (2020): DATASET.xlsx. figshare. https://figshare.com/articles/_/12081405.


    Articles from PLoS ONE are provided here courtesy of PLOS

    RESOURCES