Using a cappella to explain speech and music specialization

Using a cappella to explain speech and music specialization
Figure shows original song (bottom left) and its spectrogram (above it, in blue). This spectrogram can be decomposed according to the amount of energy contained in spectral and temporal modulation rates (central panel). Auditory cortex on the right and left sides of the brain (right side of figure) decode melody and speech, respectively, because the melody depends more on spectral modulations and the speech depends more on temporal modulations. Credit: Robert Zatorre

Speech and music are two fundamentally human activities that are decoded in different brain hemispheres. A new study used a unique approach to reveal why this specialization exists.

Researchers at The Neuro (Montreal Neurological Institute-Hospital) of McGill University created 100 a capella recordings, each of a soprano singing a sentence. They then distorted the recordings along two fundamental auditory dimensions: spectral and temporal dynamics, and had 49 participants distinguish the words or the melodies of each song. The experiment was conducted in two groups of English and French speakers to enhance reproducibility and generalizability. The experiment is demonstrated here:

They found that for both languages, when the temporal information was distorted, participants had trouble distinguishing the content, but not the . Conversely, when spectral information was distorted, they had trouble distinguishing the melody, but not the speech. This shows that speech and melody depend on different acoustical features.

To test how the brain responds to these different sound features, the participants were then scanned with functional magnetic resonance imaging (fMRI) while they distinguished the sounds. The researchers found that speech processing occurred in the left , while melodic processing occurred in the right auditory cortex.

A new study used a unique approach to reveal why music and speech are decoded in different brain hemispheres. Credit: Anthony Revoy, Shawn Hayward

Music and speech exploit different ends of the spectro-temporal continuum

Next, they set out to test how degradation in each acoustic dimension would affect brain activity. They found that degradation of the spectral dimension only affected activity in the right auditory cortex, and only during melody perception, while degradation of the temporal dimension affected only the left auditory cortex, and only during speech perception. This shows that the differential response in each hemisphere depends on the type of acoustical information in the stimulus.

Previous studies in animals have found that neurons in the auditory respond to particular combinations of spectral and temporal energy, and are highly tuned to sounds that are relevant to the animal in its natural environment, such as communication sounds. For humans, both speech and are important means of communication. This study shows that music and speech exploit different ends of the spectro-temporal continuum, and that hemispheric specialization may be the 's way of optimizing the processing of these two communication methods.

Solving the mystery of hemispheric specialization

"It has been known for decades that the two hemispheres respond to speech and music differently, but the physiological basis for this difference remained a mystery," says Philippe Albouy, the study's first author. "Here we show that this hemispheric specialization is linked to basic acoustical features that are relevant for speech and music, thus tying the finding to basic knowledge of neural organization."

The results were published in the journal Science on Feb. 28, 2020.

More information: P. Albouy el al., "Distinct sensitivity to spectrotemporal modulation supports brain asymmetry for speech and melody," Science (2020). … 1126/science.aaz3468

"Splitting speech and music," Science (2020). … 1126/science.aba7913

Journal information: Science

Provided by McGill University
Citation: Using a cappella to explain speech and music specialization (2020, February 27) retrieved 28 September 2023 from
This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no part may be reproduced without the written permission. The content is provided for information purposes only.

Explore further

How the brain distinguishes between voice and sound


Feedback to editors