Abstract
Speech is a defining human behavior, and this ability depends critically on speech motor cortex. While the ventral precentral and postcentral gyri are classically regarded as chiefly articulatory and somatosensory regions, a growing body of literature challenges this simplification. Most prior research, however, has examined cued or structured speech production tasks, neglecting the automatic, overlearned speech commonly utilized in clinical assessment. Consequently, the neural dynamics and precise timing of cortical recruitment during automatic speech remain poorly understood. Here, we present intracranial electrocorticography (ECoG) recordings from the left perisylvian cortex in participants performing automatic speech such as counting and recitation of overlearned sequences. We investigate neural dynamics using encoding (multivariate temporal response function) and decoding (deep neural network speech synthesis) models. We show that automatic speech engages a distributed network across superior temporal, precentral, and post-central cortices, characterized by attenuated pre-articulatory activity and weaker frontal encoding. Furthermore, two complementary decoding strategies reveal that speech motor cortex represents a mixture of feedforward and feedback signals, with a subset of sites exhibiting exclusively feed-forward dynamics. These results delineate the spatiotemporal cortical organization of automatic speech and establish that the speech motor cortex supports more complex dynamics than purely feedforward control.
Full Text Availability
The license terms selected by the author(s) for this preprint version do not permit archiving in PMC. The full text is available from the preprint server.
