Montreal Neurological Institute, McGill University, Montreal, Quebec, Canada.
PLoS Biol. 2012;10(7):e1001372. doi: 10.1371/journal.pbio.1001372. Epub 2012 Jul 31.
Music and speech are often cited as characteristically human forms of communication. Both share the features of hierarchical structure, complex sound systems, and sensorimotor sequencing demands, and both are used to convey and influence emotions, among other functions [1]. Both music and speech also prominently use acoustical frequency modulations, perceived as variations in pitch, as part of their communicative repertoire. Given these similarities, and the fact that pitch perception and production involve the same peripheral transduction system (cochlea) and the same production mechanism (vocal tract), it might be natural to assume that pitch processing in speech and music would also depend on the same underlying cognitive and neural mechanisms. In this essay we argue that the processing of pitch information differs significantly for speech and music; specifically, we suggest that there are two pitch-related processing systems, one for more coarse-grained, approximate analysis and one for more fine-grained accurate representation, and that the latter is unique to music. More broadly, this dissociation offers clues about the interface between sensory and motor systems, and highlights the idea that multiple processing streams are a ubiquitous feature of neuro-cognitive architectures.
音乐和言语常被认为是人类特有的两种交流形式。它们都具有层次结构、复杂的音响系统和运动感觉序列需求的特点,都被用于传达和影响情感等功能[1]。音乐和言语也都显著地使用声学频率调制,即音高的变化,作为其交流手段的一部分。鉴于这些相似性,以及音高感知和产生涉及相同的外围转导系统(耳蜗)和相同的产生机制(声道)的事实,人们可能会自然地认为,言语和音乐中的音高处理也依赖于相同的潜在认知和神经机制。在本文中,我们认为音高信息的处理在言语和音乐中存在显著差异;具体而言,我们提出存在两种与音高相关的处理系统,一种用于更粗糙、近似的分析,另一种用于更精细、准确的表示,而后者是音乐所特有的。更广泛地说,这种分离为感官和运动系统之间的界面提供了线索,并强调了多个处理流是神经认知架构的普遍特征这一观点。