Department of Speech, Language, and Hearing Sciences, Boston University, Boston, MA 02215
Graduate Program for Neuroscience, Boston University, Boston, MA 02215.
eNeuro. 2023 Oct 6;10(10). doi: 10.1523/ENEURO.0474-22.2023. Print 2023 Oct.
Spoken language contains information at a broad range of timescales, from phonetic distinctions on the order of milliseconds to semantic contexts which shift over seconds to minutes. It is not well understood how the brain's speech production systems combine features at these timescales into a coherent vocal output. We investigated the spatial and temporal representations in cerebral cortex of three phonological units with different durations: consonants, vowels, and syllables. Electrocorticography (ECoG) recordings were obtained from five participants while speaking single syllables. We developed a novel clustering and Kalman filter-based trend analysis procedure to sort electrodes into temporal response profiles. A linear discriminant classifier was used to determine how strongly each electrode's response encoded phonological features. We found distinct time-courses of encoding phonological units depending on their duration: consonants were represented more during speech preparation, vowels were represented evenly throughout trials, and syllables during production. Locations of strongly speech-encoding electrodes (the top 30% of electrodes) likewise depended on phonological element duration, with consonant-encoding electrodes left-lateralized, vowel-encoding hemispherically balanced, and syllable-encoding right-lateralized. The lateralization of speech-encoding electrodes depended on onset time, with electrodes active before or after speech production favoring left hemisphere and those active during speech favoring the right. Single-electrode speech classification revealed cortical areas with preferential encoding of particular phonemic elements, including consonant encoding in the left precentral and postcentral gyri and syllable encoding in the right middle frontal gyrus. Our findings support neurolinguistic theories of left hemisphere specialization for processing short-timescale linguistic units and right hemisphere processing of longer-duration units.
口语包含了在多个时间尺度上的信息,从毫秒级的语音区别到数秒到数分钟变化的语义语境。大脑的言语产生系统如何将这些时间尺度上的特征组合成连贯的声音输出,目前还不是很清楚。我们研究了大脑皮层中三个具有不同持续时间的语音单元的空间和时间表示:辅音、元音和音节。我们从五名参与者那里获取了说话时单个音节的脑电描记图(ECoG)记录。我们开发了一种新的聚类和基于卡尔曼滤波的趋势分析程序,将电极分类为时间响应分布。线性判别分类器用于确定每个电极的响应对语音特征的编码强度。我们发现,根据其持续时间,不同的语音单元有不同的编码时间进程:辅音在言语准备期间表现更为明显,元音在整个试验过程中均匀表现,而音节在产生过程中表现。强烈编码语音的电极位置(前 30%的电极)同样取决于语音元素的持续时间,辅音编码电极偏向左侧,元音编码电极在半球上平衡,而音节编码电极偏向右侧。编码语音的电极的偏侧化取决于起始时间,在言语产生之前或之后活跃的电极偏向左侧,而在言语过程中活跃的电极偏向右侧。单电极语音分类揭示了具有特定语音元素优先编码的皮质区域,包括左侧中央前回和中央后回的辅音编码以及右侧额中回的音节编码。我们的研究结果支持了左半球专门处理短时间尺度语言单元、右半球处理长时间尺度单元的神经语言学理论。