Skip to main content
Proceedings of the National Academy of Sciences of the United States of America logoLink to Proceedings of the National Academy of Sciences of the United States of America
. 2014 Mar 31;111(16):5837–5841. doi: 10.1073/pnas.1318261111

Language universals at birth

David Maximiliano Gómez a,b,1, Iris Berent c, Silvia Benavides-Varela a,d,e, Ricardo A H Bion f, Luigi Cattarossi g, Marina Nespor a, Jacques Mehler a
PMCID: PMC4000834  PMID: 24706790

Significance

It is well known that across languages, certain structures are preferred to others. For example, syllables like blif are preferred to syllables like bdif and lbif. But whether such regularities reflect strictly historical processes, production pressures, or universal linguistic principles is a matter of much debate. To address this question, we examined whether some precursors of these preferences are already present early in life. The brain responses of newborns show that, despite having little to no linguistic experience, they reacted to syllables like blif, bdif, and lbif in a manner consistent with adults’ patterns of preferences. We conjecture that this early—possibly universal—bias helps shaping language acquisition.

Keywords: human newborns, speech perception, NIRS, sonority, phonology

Abstract

The evolution of human languages is driven both by primitive biases present in the human sensorimotor systems and by cultural transmission among speakers. However, whether the design of the language faculty is further shaped by linguistic biological biases remains controversial. To address this question, we used near-infrared spectroscopy to examine whether the brain activity of neonates is sensitive to a putatively universal phonological constraint. Across languages, syllables like blif are preferred to both lbif and bdif. Newborn infants (2–5 d old) listening to these three types of syllables displayed distinct hemodynamic responses in temporal-perisylvian areas of their left hemisphere. Moreover, the oxyhemoglobin concentration changes elicited by a syllable type mirrored both the degree of its preference across languages and behavioral linguistic preferences documented experimentally in adulthood. These findings suggest that humans possess early, experience-independent, linguistic biases concerning syllable structure that shape language perception and acquisition.


It is well known that the design of human language is shaped by both cultural and biological constraints (1, 2). However, whether those biological constraints are limited to sensorimotor restrictions on the production and perception of language, or whether they also include linguistic principles, remains controversial. To address this question, we examine the sensitivity of newborn infants to a putatively universal linguistic principle that defines syllabic structure. Our results suggest that precursors of this universal principle are active in the newborn brain. These findings are consistent with the presence of biological linguistic constraints on language acquisition.

Phonology is pivotal to the design of the human language faculty (36). Not only is it present in every language—spoken or signed (7)—but distinct languages appear to share common phonological restrictions. One such restriction concerns the internal structure of syllables. Across languages, syllables like blif are more frequent than syllables like lbif (8, 9). These restrictions are attributed to a putatively universal constraint on syllable structure, known as the Sonority Sequencing Principle (SSP; ref. 10). Sonority (s) is a scalar phonological property that correlates with the salience of phonological elements (e.g., loudness of speech sounds; ref. 11). Least sonorous are obstruents (e.g., /t/, /b/, /f/), with a sonority level of 1 (s = 1), followed by nasals (e.g., /m/, /n/, s = 2), liquids (e.g., /l/, /r/, s = 3), glides (/w/, /j/, s = 4), and finally vowels, which are the most sonorant phonemes of all (s = 5). The SSP states that syllables maximize the sonority distance (∆s) from their margins to their nucleus—the larger the sonority distance, the better formed the syllable is (10). In syllables like blif, there is a rise in sonority from the obstruent (b) to the liquid (l) (∆s = 2), whereas in lbif, there is a sonority fall (∆s = −2). In between these two extremes are syllables such as bdif, where the two initial consonants (two obstruents) exhibit a sonority plateau (∆s = 0). Under the hypothesis that languages favor a large distance in sonority in onset position, syllables like blif (∆s = 2) are expected to be better formed than bdif (∆s = 0), which, in turn, are expected to be better formed than lbif (∆s = −2). The frequency of those syllables across languages is consistent with this prediction (12). Not only are syllables like lbif infrequent—they are less frequent than either bdif or blif—but languages that tolerate such syllables tend to also exhibit their better-formed counterparts (i.e., bdif, blif) (8, 12).

Experimental research has further shown that the SSP modulates the perception of syllables by individual speakers. As the syllable structure becomes worse formed (as defined by the SSP), people tend to systematically misidentify the syllable. For example, people tend to misidentify syllables such as lbif (e.g., as lebif), they are less likely to misidentify bdif, and least likely to misidentify blif. This phenomenon has been documented in numerous languages with both isolated syllables [e.g., English (1214), French (15), Hebrew (16), Korean (17)] and continuous artificial speech (18). Crucially, these effects occur even if the specific syllables under investigation are unattested in the native language of participants (1218). These findings open up the possibility that the SSP might not be induced from linguistic experience. Nonetheless, these results are silent as to how the SSP arises in the course of development—whether such restrictions are active at birth, or whether their emergence is triggered by experience with the articulatory production of spoken language.

Human newborns provide a superb opportunity to investigate the ontogenetic origin of biases like the SSP*. Hearing newborns process speech preferentially with the left hemisphere (21), distinguish between languages based on their rhythmic properties (22), learn precociously the properties of intonation of their maternal language (23), and discriminate phonemic changes in syllables despite speaker variability (24). Moreover, by 4 mo of age, newborns exhibit phonological processing in left temporal brain areas, as evidenced by their enhanced sensitivity to phonetic changes that cross phonemic boundaries (25). These early dispositions suggest that some phonological knowledge is already in place in the first days after birth. Here, we asked whether newborn brains display precursors of the SSP. More specifically, we measured hemodynamic activity of healthy newborn infants during passive listening of speech items by using functional near-infrared spectroscopy (NIRS; refs. 2628). Of interest is whether their hemodynamic responses reflect the syllable preferences predicted by the SSP.

Results and Discussion

Experiment 1 explored whether the brains of newborns react differentially to syllables that are well- or extremely ill-formed, as defined by the SSP. Twenty-four healthy newborns (mean age = 2.9 d, SD = 0.83) listened to blocks of C1C2VC3 (C: consonant, V: vowel) syllables displaying either a sonority rise or a sonority fall between C1 and C2 (e.g., blif and lbif, respectively; see Table S1 for the full list of syllables). Syllables in both conditions were statistically undistinguishable in terms of average pitch, intensity, duration, and number of feature changes between C1 and C2. Because of their role in neonatal speech perception (21), we concentrated our analysis on bilateral temporal-perisylvian areas. Additionally, our NIRS probes allowed us to record bilateral frontoparietal cortex, which has been linked to the processing of suprasegmental properties of speech (29) (Fig. 1). Fig. 2A presents results for both oxyhemoglobin and deoxyhemoglobin, and all regions of interest. We found significant differences between conditions in oxyhemoglobin concentrations in both left temporal [t(22) = 2.70, P = 0.04] and right frontoparietal [t(21) = 2.97, P = 0.029] areas (all P values are corrected according to the Holm–Bonferroni procedure; see also Table S2 for all statistics including uncorrected P values). In both areas, well-formed syllables elicited lower oxyhemoglobin concentrations than ill-formed syllables. As in previous research with newborn infants (27), deoxyhemoglobin concentrations did not differ significantly between conditions (all uncorrected P values >0.33).

Fig. 1.

Fig. 1.

NIRS probes were positioned on the heads of newborns so as to cover two regions of interest (ROIs) on each brain hemisphere. The inferior ROIs correspond roughly to temporal lobe and perisylvian areas, whereas the superior ROIs correspond to frontoparietal areas. Because the NIRS probes have a fixed size, the variability in head size and shape of newborns prevents precise scalp/brain mappings.

Fig. 2.

Fig. 2.

Differences in concentration changes (ill-formed minus well-formed) for both hemoglobin species and all four ROIs. (A) Experiment 1 (n = 24), contrasting sonority rises and falls (e.g., blif vs. lbif). (B) Experiment 2 (n = 24), contrasting sonority rises and plateaus (e.g., blif vs. bdif). *P < 0.05 (Holm-Bonferroni corrected).

These results show that newborn brains distinguish syllables that obey the SSP from syllables that violate it. To do so, newborns must have abstracted a common structure from each syllable type, notwithstanding huge phonetic and phonemic variations. The difference in activity in left temporal cortex was expected given the strong evidence linking early phonological capacities to this brain area (21, 25). Frontoparietal differences in the right hemisphere could be due to differences in suprasegmental properties of speech such as the speech envelope (29, 30), or to frontal mechanisms monitoring salient stimuli (31).

Experiment 2 investigated sensitivity of newborns to an additional syllable type. A new group of 24 newborns (mean age = 3.0 d, SD = 0.66) listened to blocks of C1C2VC3 syllables that presented either a sonority rise (e.g., blif) or a sonority plateau (e.g., bdif, where C1 and C2 have the same sonority) in their onset. Across languages, onsets of level sonority are dispreferred to sonority rises (12), and similar preferences have been documented behaviorally among adult individuals despite no experience with either structure (17). We thus ask whether the distinction between sonority rise and plateau is also present at birth and, if so, whether its hemodynamic manifestation aligns with the results of Experiment 1. Beyond its theoretical significance, the comparison of sonority plateaus and rises is convenient because it allows us to control for some acoustic and phonetic correlates of syllable structure present in Experiment 1. Indeed, the material therein used can be discriminated on the basis of their sonority contour, its speech envelope, or the fact that all syllables started with obstruents in one condition and with liquids in the other. Experiment 2 attenuates these factors and, thus, provides a stringent test for the sonority account.

Results revealed that oxyhemoglobin concentrations elicited by well-formed syllables are significantly lower than concentrations elicited by plateaus in the left temporal cortex (t(22) = 2.88, P = 0.035), in agreement with the sonority account and the results of Experiment 1. However, in the present experiment, discrimination was not evident at the right frontoparietal cortex [t(17) = 1.86, P = 0.16; see Fig. 2B]. Effect size measured by Cohen’s d dropped 30% compared with Experiment 1 in the right frontoparietal region. This drop could be either due to the attenuation of phonetic cues in this experiment, or to the fact that the ill-formed syllables included in Experiment 2 (e.g., bdif) had a larger sonority distance than those in Experiment 1 (e.g., lbif).

Across Experiments 1 and 2, ill-formed syllables (as defined by the SSP) elicited higher oxyhemoglobin concentrations in left temporal-perisylvian areas. Specifically, compared with the well-formed sonority rises (e.g., blif), oxyhemoglobin concentrations were higher for sonority falls (e.g., lbif, in Experiment 1) and for sonority plateaus (e.g., bdif, in Experiment 2). Moreover, inspection of hemodynamic responses to each syllable type separately (Fig. 3) indicates that responses to well- and ill-formed syllables differ qualitatively. Well-formed syllables produce negative variations in oxyhemoglobin levels, a pattern absent for ill-formed syllables. Together, these results suggest that neonates’ brain activity systematically distinguishes well-formed syllables from ill-formed ones. These findings suggest that a precursor of the adult linguistic preference might be active at birth.

Fig. 3.

Fig. 3.

Hemodynamic responses in the left inferior region of interest for Experiments 1 (Upper) and 2 (Lower). Left correspond to well-formed syllables (e.g., blif), whereas Right depict ill-formed syllables (e.g., lbif, bdif). Black lines depict baseline level and the time of stimulation. For all graphs, x axes display time elapsed from stimulation onset in seconds, and y axes display variations in hemoglobin concentrations measured in mmol•mm. Error lines indicate SEM.

To further determine whether newborns’ response to bl and lb clusters depends on their syllable position, we next measured responses to the same clusters located across syllables. In Experiment S1 (SI Results and Discussion), we added a single vowel at the beginning of an ill-formed syllable like lbif to obtain disyllabic words like olbif. Because the sonority fall now spans across two syllables (ol.bif), rather than a syllable onset (e.g., lbif), such words should be perfectly well-formed. In line with this prediction, our results show that newborns’ brain responses to disyllables like oblif and olbif do not differ. Finding that the same clusters (lb vs. bl) elicit markedly different responses, depending on their syllable position (e.g., lbif vs. olbif), suggests that infants constrain syllable structure rather than consonant sequencing per se.

General Discussion

These results show that neonates are sensitive to putatively universal restrictions on syllable structure. The observation of such regularities close to birth, before the onset of experience with language production, shows that sonority-related biases in humans do not require extensive linguistic experience (34) or ample practice with language production. Although we cannot presently rule out the possibility of learning from prenatal perceptual experience (i.e., from Italian—the language spoken by the infants’ mothers), this possibility is exceedingly unlikely given the strong intrauterine attenuation of sound frequencies above 300 Hz (35, 36)—conditions that render consonant discrimination improbable. Moreover, the proposal that infants can learn the syllable structure of their language in utero fails to explain the evidence for the acquisition of the phonotactic patterns of their language only by the end of the first year of life, and not earlier (37, 38).

Taken as a whole, our results show that human newborns listening to spoken syllables exhibit precursors of universal linguistic preferences. Our present findings are moot with respect to the precise source of such preferences—whether they reflect grammatical phonological constraints or phonetic pressures. Nonetheless, newborns display distinct oxyhemoglobin responses for well- and ill-formed syllables in channels located over their left temporal cortex. We consistently observed that oxyhemoglobin curves were higher for classes of syllables that are dispreferred across languages, in agreement with previous research showing that the left temporal cortex is sensitive to phonological constraints very early in development. These results suggest that a precursor of the universal sonority-related preferences seen in adults (1218) is present close to birth.

Biological biases have been shown to guide cultural transmission of birdsong: isolated colonies of zebra finches converge to a song repertoire similar to their wild type (39). We propose that subtle phonological biases such as the one here documented guide the cultural evolution of languages via similar mechanisms.

Materials and Methods

Participants.

Our samples for Experiments 1, 2, and S1 are composed by 24 healthy newborns each. Newborns were recruited and tested at the Hospital, Azienda Ospedaliera Santa Maria della Misericordia, in Udine, Italy. Newborns were considered eligible to participate if they had a head circumference of at least 33 cm, an Apgar score of at least 7 at the first minute, and no cefalhematoma. Parents signed consent forms at the beginning of the experimental session. Our procedures and protocols were approved by the Scuola Internazionale Superiore di Studi Avanzati Ethics Committee.

Eight boys and 16 girls (mean age = 2.9 d, SD = 0.83) participated in Experiment 1. They had Apgar scores of 8.5 ± 0.83 and 9.0 ± 0.20 at the first and fifth minute, respectively, a gestational age of 39.3 ± 1.17 wk, weighted 3.310 ± 0.307 kg at birth, and had a head circumference of 34.5 ± 1.0 cm. An additional eight newborns were tested but rejected because of crying or fussiness (n = 4), difficulties in obtaining good NIRS signal (n = 3), or experimental error (n = 1).

Fourteen boys and 10 girls (mean age = 3.0 d, SD = 0.66) participated in Experiment 2. They had Apgar scores of 8.5 ± 0.51 and 9.3 ± 0.44 at the first and fifth minute, respectively, a gestational age of 38.9 ± 1.50 wk, weighted 3.429 ± 0.378 kg at birth, and had a head circumference of 35.0 ± 0.8 cm. An additional 16 newborns were tested but rejected due to crying or fuzziness (n = 10) or difficulties in obtaining good NIRS signal (n = 6).

Ten boys and 14 girls (mean age = 3.2 d, SD = 0.80) participated in Experiment S1. They had Apgar scores of 8.6 ± 0.7 and 9.1 ± 0.5 at the first and fifth minute, respectively, a gestational age of 39.3 ± 1.3 wk, weighted 3.359 ± 0.398 kg at birth, and had a head circumference of 34.7 ± 1.1 cm. An additional 18 newborns were tested but rejected because of crying or fuzziness (n = 11) or difficulties in obtaining good NIRS signal (n = 7).

Stimuli.

A female native speaker of Russian recorded all words used in our experiments. Russian allows all syllable types studied in our experiment. Thus, Russian speakers can produce those items natively. A large list of words was recorded in a single session in a sound-attenuated booth, from which a phonologist selected sets of eight words per condition among the best recorded exemplars. We used CCVC (C, consonant; V, vowel) syllables for Experiments 1 and 2 and VCCVC bisyllables for Experiment S1 (Table S1). Within each experiment, syllable sets did not statistically differ in terms of their duration, average pitch, or number of feature changes between the two adjacent consonants (t tests, all P values >0.29). Sound intensity was set to 70 dB for all words.

Procedure.

Newborns were tested in their crib in a silent room assisted by a medical doctor, either during sleep or in a quiet state of alert. Sound stimuli were presented via two loudspeakers located approximately 70 cm in front of the neonate. A Mac PowerPC G5 operated the fNIRS machine and presented the auditory stimuli using the software PsyScope X (http://psy.ck.sissa.it). Our procedure followed a block design (40): newborns listened to blocks consisting of words from a given type. Each stimulation block lasted approximately 13 s, presenting all eight words of a given condition in random order. Random pauses separated consecutive words (0.5 or 1.5 s) and blocks (25 or 35 s). We presented 10 blocks per condition, for a total duration of the experimental session of approximately 15 min. An infrared videocamera was used to monitor the state and behavior of neonates.

Data Acquisition.

The cortical hemodynamic activity of newborns was recorded by using fNIRS (ETG-4000; Hitachi). This machine emits continuous near-infrared light of two wavelengths (695 and 830 nm) through 10 emitters and records light absorption through eight detectors. Emitters and detectors are arrayed in two Chevron-shaped silicon probes, providing 24 recording channels. Emitter-detector distance is 3 cm, sampling rate is 10 Hz, and total laser power output per optical fiber was set to 0.75 mW. Using the vertex and tragus as skull landmarks, we positioned probes over the scalps of newborns: one over each hemisphere, with the concave side of the Chevron shape surrounding the ears.

Data Processing and Analysis.

Variations in oxyhemoglobin and deoxyhemoglobin were obtained from light absorption recordings by using the modified Beer–Lambert Law (27). Hemodynamic signals were band-pass filtered between 0.02 and 0.5 Hz. Epochs were extracted starting 5 s before each block and finishing 15 s after its end, for a total epoch length of 33 s. Single channels for specific blocks were rejected if light absorption was less than 1% of the total light emitted or if the hemodynamic signal contained movement artifacts (changes in the signal greater than 0.1 mmol•mm in an interval of 0.2 s). Epochs with more than 12 rejected channels were excluded. Only participants with at least three good epochs per condition were considered for analysis.

A baseline was linearly fitted between the mean of the initial and final 5 s of each epoch and subtracted from the signal. Statistical analyses involved the area under each hemodynamic curve over the whole epoch: Area values were averaged for each newborn across all channels composing each region of interest (ROI). Comparisons were conducted by means of paired t tests, using the Holm–Bonferroni procedure (41) to correct for multiple comparisons. Because rejection criteria were applied on a channel-per-channel basis separately for each infant, t tests corresponding to different ROIs might have different numbers of degrees of freedom.

Temporal-perisylvian ROIs comprised channels above the ear: 11 and 12 (left hemisphere) and 23 and 24 (right hemisphere). Frontoparietal ROIs comprised channels closest to the vertex: 1 and 2 (left hemisphere) and 13 and 14 (right hemisphere). Given the great variability in the head shapes of neonates and the fixed shape of the NIRS probes, these localizations are approximate.

Supplementary Material

Supporting Information

Acknowledgments

We thank the staff at the Neonatology and Obstetrics Departments of Udine Hospital for their invaluable assistance and to the parents of our newborn subjects for their participation. The research leading to these results has received funding from the European Research Council under the European Union’s Seventh Framework Programme (FP7/2007-2013)/European Research Council Grant Agreement 269502 (PASCAL) (to J.M.).

Footnotes

The authors declare no conflict of interest.

This article is a PNAS Direct Submission.

*Some researchers have argued that sonority is mostly a phonetic—as opposed to phonological—construct (19, 20). Our present results do not speak to this debate, because we have no basis to determine whether responses of infants reflect phonetic or phonological preferences. A purely phonetic account, however, fails to capture the adult data (e.g., the emergence of sonority effects with printed materials; ref. 14), suggesting that sonority effects might be partly phonological.

Negative oxyhemoglobin deflections, or more generally noncanonical hemodynamic responses, are relatively frequent in infant research. Cases have been documented for “visual, olfactory, sensory-motor, and auditory cortices” (32). Although there is not yet consensus on the causes of such responses, some investigations suggest a role of systemic vascular activity (32, 33). Systemic activity cannot fully account for our results, because the data suggest that the crucial differences are occurring in left temporal-perisylvian brain areas. Even if systemic activity could display such localization, this activity must reflect the linguistic structure of the stimuli, because well-formed stimuli elicit different responses than ill-formed ones. Regardless of origin—systemic or cortical—such changes thus demonstrate the presence of a linguistic bias on syllable structure in newborn infants.

Our hemodynamic data do not directly indicate preference on the part of newborns. Nonetheless, there is a striking resemblance between the pattern of results across our experiments and the well-documented cross-linguistic preferences for syllabic well-formedness. Typological data and behavior of adults (e.g., ref. 12) show that syllables with sonority rises (blif) are preferred to both falls (lbif) and plateaus (bdif). Our findings closely match this pattern, inasmuch as well-formed sequences produce systematically higher hemodynamic responses than ill-formed sequences. Whether the absolute strength of hemodynamic response indicates preference is unknown. However, for our argument, the critical factor is not the absolute direction or magnitude (e.g., “stronger response = preference”) but rather its pattern. Across our experiments, different types of ill-formed stimuli consistently produce the same pattern of hemodynamic response relative to well-formed stimuli. The consistency of this pattern is significant, because it suggests that neonates lump our different ill-formed stimuli as a category that is distinct from well-formed stimuli, just as adults do.

This article contains supporting information online at www.pnas.org/lookup/suppl/doi:10.1073/pnas.1318261111/-/DCSupplemental.

References

  • 1.Chomsky N. Three factors in language design. Linguist Inq. 2005;36(1):1–22. [Google Scholar]
  • 2.Fitch WT. Unity and diversity in human language. Philos Trans R Soc Lond B Biol Sci. 2011;366(1536):376–388. doi: 10.1098/rstb.2010.0223. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Kenstowicz M. Phonology in Generative Grammar. Oxford: Blackwell Publications; 1994. [Google Scholar]
  • 4.Chomsky N, Lasnik H. 1993. The theory of principles and parameters. Syntax: An International Handbook of Contemporary Research, eds Jacobs J, von Stechow A, Sternefeld W, Vennemann T (de Gruyter, Berlin), Vol 1, pp 506–569.
  • 5.Nespor M, Vogel I. 2007. Prosodic Phonology (de Gruyter, Berlin)
  • 6.Berent I. The Phonological Mind. Cambridge, UK: Cambridge Univ Press; 2013. [Google Scholar]
  • 7.Perlmutter DM. Sonority and syllable structure in American Sign Language. Linguist Inq. 1992;23(3):407–442. [Google Scholar]
  • 8.Greenberg JH. Some generalizations concerning initial and final consonant clusters. In: Greenberg JH, Ferguson CA, Moravcsik EA, editors. Universals of Human Language. Vol 2. Stanford: Stanford Univ Press; 1978. pp. 243–279. [Google Scholar]
  • 9.Kreitman R. 2012. On the relations between [sonorant] and [voice]. Consonant Clusters and Structural Complexity, eds Hoole P, Bombien L, Pouplier M, Mooshammer C, Kühnert B (de Gruyter, Berlin), pp 33–70.
  • 10.Clements GN. The role of the sonority cycle in core syllabification. In: Kingston J, Beckman ME, editors. Papers in Laboratory Phonology I: Between the Grammar and the Physics of Speech. New York: Cambridge Univ Press; 1990. pp. 283–333. [Google Scholar]
  • 11.Parker S. Sound level protrusions as physical correlates of sonority. J Phonetics. 2008;36(1):55–90. [Google Scholar]
  • 12.Berent I, Steriade D, Lennertz T, Vaknin V. What we know about what we have never heard: Evidence from perceptual illusions. Cognition. 2007;104(3):591–630. doi: 10.1016/j.cognition.2006.05.015. [DOI] [PubMed] [Google Scholar]
  • 13.Berent I, Harder K, Lennertz T. Phonological universals in early childhood: Evidence from sonority restrictions. Lang Acquis. 2011;18(4):281–293. doi: 10.1080/10489223.2011.580676. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Berent I, Lennertz T, Smolensky P, Vaknin-Nusbaum V. Listeners’ knowledge of phonological universals: Evidence from nasal clusters. Phonology. 2009;26(1):75–108. doi: 10.1017/S0952675709001729. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Maïonchi-Pino N, de Cara B, Écalle J, Magnan A. Are French dyslexic children sensitive to consonant sonority in segmentation strategies? Preliminary evidence from a letter detection task. Res Dev Disabil. 2012;33(1):12–23. doi: 10.1016/j.ridd.2011.07.045. [DOI] [PubMed] [Google Scholar]
  • 16.Berent I, Vaknin-Nusbaum V, Balaban E, Galaburda AM. Phonological generalizations in dyslexia: The phonological grammar may not be impaired. Cogn Neuropsychol. 2013;30(5):285–310. doi: 10.1080/02643294.2013.863182. [DOI] [PubMed] [Google Scholar]
  • 17.Berent I, Lennertz T, Jun J, Moreno MA, Smolensky P. Language universals in human brains. Proc Natl Acad Sci USA. 2008;105(14):5321–5325. doi: 10.1073/pnas.0801469105. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Ettlinger M, Finn AS, Hudson Kam CL. The effect of sonority on word segmentation: Evidence for the use of a phonological universal. Cogn Sci. 2012;36(4):655–673. doi: 10.1111/j.1551-6709.2011.01211.x. [DOI] [PubMed] [Google Scholar]
  • 19.Wright RA. A review of perceptual cues and cue robustness. In: Hayes B, Kirchner R, Steriade D, editors. Phonetically Based Phonology. New York: Cambridge Univ Press; 2004. pp. 34–57. [Google Scholar]
  • 20.Davidson L. Phonetic bases of similarities in cross-language production: Evidence from English and Catalan. J Phonetics. 2010;38(2):272–288. [Google Scholar]
  • 21.Peña M, et al. Sounds and silence: An optical topography study of language recognition at birth. Proc Natl Acad Sci USA. 2003;100(20):11702–11705. doi: 10.1073/pnas.1934290100. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.Nazzi T, Bertoncini J, Mehler J. Language discrimination by newborns: Towards an understanding of the role of rhythm. J Exp Psychol Hum Percept Perform. 1998;24(3):756–766. doi: 10.1037//0096-1523.24.3.756. [DOI] [PubMed] [Google Scholar]
  • 23.Mampe B, Friederici AD, Christophe A, Wermke K. Newborns’ cry melody is shaped by their native language. Curr Biol. 2009;19(23):1994–1997. doi: 10.1016/j.cub.2009.09.064. [DOI] [PubMed] [Google Scholar]
  • 24.Dehaene-Lambertz G, Peña M. Electrophysiological evidence for automatic phonetic processing in neonates. Neuroreport. 2001;12(14):3155–3158. doi: 10.1097/00001756-200110080-00034. [DOI] [PubMed] [Google Scholar]
  • 25.Dehaene-Lambertz G, Baillet S. A phonological representation in the infant brain. Neuroreport. 1998;9(8):1885–1888. doi: 10.1097/00001756-199806010-00040. [DOI] [PubMed] [Google Scholar]
  • 26.Lloyd-Fox S, Blasi A, Elwell CE. Illuminating the developing brain: The past, present, and future of functional near infrared spectroscopy. Neurosci Biobehav Rev. 2010;34(3):269–284. doi: 10.1016/j.neubiorev.2009.07.008. [DOI] [PubMed] [Google Scholar]
  • 27.Gervain J, et al. Near-infrared spectroscopy: A report from the McDonnell infant methodology consortium. Dev Cogn Neurosci. 2011;1(1):22–46. doi: 10.1016/j.dcn.2010.07.004. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.Rossi S, Telkemeyer S, Wartenburger I, Obrig H. Shedding light on words and sentences: Near-infrared spectroscopy in language research. Brain Lang. 2012;121(2):152–163. doi: 10.1016/j.bandl.2011.03.008. [DOI] [PubMed] [Google Scholar]
  • 29.Lindell AK. In your right mind: Right hemisphere contributions to language processing and production. Neuropsychol Rev. 2006;16(3):131–148. doi: 10.1007/s11065-006-9011-9. [DOI] [PubMed] [Google Scholar]
  • 30.Abrams DA, Nicol T, Zecker S, Kraus N. Right hemisphere auditory cortex is dominant for coding syllable patterns in speech. J Neurosci. 2008;28(15):3958–3965. doi: 10.1523/JNEUROSCI.0187-08.2008. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 31.Vallesi A. Organisation of executive functions: Hemispheric asymmetries. Journal of Cognitive Psychology. 2012;24(4):367–386. [Google Scholar]
  • 32.Zimmermann BB, et al. The confounding effect of systemic physiology on the hemodynamic response in newborns. In: Wolf M, et al., editors. Oxygen Transport to Tissue XXXIII. New York: Springer; 2012. pp. 103–109. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.Kirilina E, et al. The physiological origin of task-evoked systemic artefacts in functional near infrared spectroscopy. Neuroimage. 2012;61(1):70–81. doi: 10.1016/j.neuroimage.2012.02.074. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.Daland R, et al. Explaining sonority projection effects. Phonology. 2011;28(2):197–234. [Google Scholar]
  • 35.Abrams RM, et al. Fetal music perception: The role of sound transmission. Music Percept. 1998;15(3):307–317. [Google Scholar]
  • 36.Querleu D, Renard X, Versyp F, Paris-Delrue L, Crèpin G. Fetal hearing. Eur J Obstet Gynecol Reprod Biol. 1988;28(3):191–212. doi: 10.1016/0028-2243(88)90030-5. [DOI] [PubMed] [Google Scholar]
  • 37.Mazuka R, Cao Y, Dupoux E, Christophe A. The development of a phonological illusion: A cross-linguistic study with Japanese and French infants. Dev Sci. 2011;14(4):693–699. doi: 10.1111/j.1467-7687.2010.01015.x. [DOI] [PubMed] [Google Scholar]
  • 38.Jusczyk PW, Friederici AD, Wessels JMI, Svenkerud VY, Jusczyk AM. Infants’ sensitivity to the sound patterns of native language words. J Mem Lang. 1993;32(3):402–420. [Google Scholar]
  • 39.Fehér O, Wang H, Saar S, Mitra P, Tchernichovski O. De novo establishment of wild-type song culture in the zebra finch. Nature. 2009;459(7246):564–568. doi: 10.1038/nature07994. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 40.Benavides-Varela S, Gómez DM, Mehler J. Studying neonates’ language and memory capacities with functional near-infrared spectroscopy. Front Psychol. 2011;2:64. doi: 10.3389/fpsyg.2011.00064. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 41.Holm S. A simple sequentially rejective multiple test procedure. Scand J Stat. 1979;6(2):65–70. [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Supporting Information

Articles from Proceedings of the National Academy of Sciences of the United States of America are provided here courtesy of National Academy of Sciences

RESOURCES