This article has been reviewed according to Science X's editorial process and policies. Editors have highlighted the following attributes while ensuring the content's credibility:


peer-reviewed publication

trusted source


Neural activity study shows the brain processes direct speech and its echo separately

The brain processes speech and its echo separately
Speech processing in an echoic environment. Credit: Jiaxin Gao (CC-BY 4.0,

Echoes can make speech harder to understand, and tuning out echoes in an audio recording is a notoriously difficulty engineering problem. The human brain, however, appears to solve the problem successfully by separating the sound into direct speech and its echo, according to a study published in PLOS Biology by Jiaxin Gao from Zhejiang University, China, and colleagues.

The in online meetings and auditoriums that are not properly designed often have an echo lagging at least 100 milliseconds from the original speech. These echoes heavily distort speech, interfering with slowly varying sound features most important for understanding conversations, yet people still reliably understand echoic speech.

To better understand how the brain enables this, the authors used magnetoencephalography (MEG) to record while listened to a story with and without an echo. They compared the neural signals to two computational models: one simulating the brain 'adapting' to the echo, and another simulating the brain 'separating' the echo from the original speech.

Participants understood the story with over 95% accuracy, regardless of echo. The researchers observed that cortical activity tracks energy changes related to direct speech, despite the strong interference of the echo. Simulating neural adaptation only partially captured the they observed—neural activity was better explained by a model that split original speech and its echo into separate processing streams.

This remained true even when participants were told to direct their attention toward a silent film and ignore the story, suggesting that top-down attention isn't required to mentally separate direct speech and its echo.

The researchers state that auditory stream segregation may be important both for singling out a specific speaker in a crowded environment, and for clearly understanding an individual speaker in a reverberant space.

The authors add, "Echoes strongly distort the sound features of speech and create a challenge for automatic speech recognition. The , however, can segregate speech from its echo and achieve reliable recognition of echoic speech."

More information: Gao J, Chen H, Fang M, Ding N (2024) Original speech and its echo are segregated and separately processed in the human brain, PLoS Biology (2024). DOI: 10.1371/journal.pbio.3002498

Journal information: PLoS Biology
Citation: Neural activity study shows the brain processes direct speech and its echo separately (2024, February 15) retrieved 22 April 2024 from
This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no part may be reproduced without the written permission. The content is provided for information purposes only.

Explore further

Study unveils similarities between the auditory pathway and deep learning models for processing speech


Feedback to editors