This article has been reviewed according to Science X's editorial process and policies. Editors have highlighted the following attributes while ensuring the content's credibility:


peer-reviewed publication

trusted source


Study identifies two brain mechanisms for picking speech out of a crowd

Two brain mechanisms for picking speech out of a crowd
Example of listening to someone talking in a noisy environment. Credit: Zuckerman Institute, Columbia University (2023) (CC-BY 4.0,

Researchers led by Dr. Nima Mesgarani at Columbia University, US, report that the brain treats speech in a crowded room differently depending on how easy it is to hear, and whether we are focusing on it.

Publishing June 6th in the open access journal PLOS Biology, the study uses a combination of neural recordings and computer modeling to show that when we follow that is being drowned out by louder voices, phonetic information is encoded differently than in the opposite situation. The findings could help improve that work by isolating attended speech.

Focusing on speech in a crowded room can be difficult, especially when other voices are louder. However, amplifying all sounds equally does little to improve the ability to isolate these hard-to-hear voices, and hearing aids that try to only amplify attended speech are still too inaccurate for practical use.

In order to gain a better understanding of how speech is processed in these situations, the researchers at Columbia University recorded from electrodes implanted in the brains of people with epilepsy as they underwent . The patients were asked to attend to a single voice, which was sometimes louder than another voice ("glimpsed") and sometimes quieter ("masked").

The researchers used the neural recordings to generate predictive models of activity. The models showed that phonetic information of "glimpsed" speech was encoded in both primary and secondary auditory cortex of the brain, and that encoding of the attended speech was enhanced in the secondary cortex.

In contrast, phonetic information of "masked" speech was only encoded if it was the attended . Lastly, speech encoding occurred later for "masked" speech than for "glimpsed' speech. Because "glimpsed" and "masked" phonetic information appear to be encoded separately, focusing on deciphering only the "masked" portion of attended speech could lead to improved auditory attention-decoding systems for brain-controlled hearing aids.

Vinay Raghavan, the lead author of the study, says, "When listening to someone in a noisy place, your brain recovers what you missed when the background noise is too loud. Your brain can also catch bits of speech you aren't focused on, but only when the person you're listening to is quiet in comparison."

More information: Raghavan VS, O'Sullivan J, Bickel S, Mehta AD, Mesgarani N. Distinct neural encoding of glimpsed and masked speech in multitalker situations, PLoS Biology (2023). DOI: 10.1371/journal.pbio.3002128

Journal information: PLoS Biology

Citation: Study identifies two brain mechanisms for picking speech out of a crowd (2023, June 6) retrieved 24 September 2023 from
This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no part may be reproduced without the written permission. The content is provided for information purposes only.

Explore further

Clear speech in the new digital era: Speaking and listening clearly to voice-AI systems


Feedback to editors