Speaking and understanding speech share the same parts of the brain

August 16, 2011, Association for Psychological Science

The brain has two big tasks related to speech: making it and understanding it. Psychologists and others who study the brain have debated whether these are really two separate tasks or whether they both use the same regions of the brain. Now, a new study, published in the August issue of Psychological Science, a journal of the Association for Psychological Science, finds that speaking and understanding speech share the same parts of the brain, with one difference: we don't need the brain regions that control the movements of lips, teeth, and so on to understand speech.

Most studies of how speech works in the focuses on comprehension. That's mostly because it's easier to image the brains of people who are listening quietly; talking makes the head move, which is a problem when you're measuring the brain. But now, the Donders Institute at the Radboud University Nijmegen, where the study was conducted, has developed technology that allows recording from a moving brain.

Laura Menenti, a Postdoctoral Research Associate at the University of Glasgow, co-wrote the paper along with Peter Hagoort of Radboud University Nijmegen and the Institute for , Sarah Gierhan and Katrien Segaert. Menenti was initially interested in how the brain produces grammatical sentences and wanted to track the process of producing a sentence in its entirety; looking not only at its grammatical structure but also at its meaning. "What made this particularly exciting to us was that no one had managed to perform such a study before, meaning that we could explore an almost completely new topic," says Menenti.

The authors used functional to measure in people who were either listening to sentences or speaking sentences. The other problem with measuring brain activity in people who are speaking is that you have to get them to say the right kind of sentence. The authors accomplished this with a picture of an action—a man strangling a woman, say—with one person colored green and one colored red to indicate their order in the sentence. This prompted people to say either "The man is strangling the woman" or "The woman is strangled by the man." (The experiments were all carried out in Dutch.)

From this, the researchers were able to tell where in the brain three different speech tasks (computing meaning, coming up with the words, and building a grammatical sentence)—were taking place. They found that the same areas were activated for each of these tasks in people who were speaking and people who were listening to sentences. However, although some studies have suggested that while people are listening to speech, they silently articulate the words in order to understand them, the authors found no involvement of motor regions when people were listening.

According to Menenti, though the study was largely designed to answer a specific theoretical question, it also points towards some useful avenues for treatment of people with language-related problems. It suggests that while it sometimes seems that people with comprehension problems may have intact production, and vice versa, this may not necessarily be the case. According to Menenti, "Our data suggest that these problems would be expected to always at least partly coincide. On the other, our data confirm the idea that many different processes in the language system, such as understanding meaning or grammar, can at least partly, be damaged independently of each other."

Related Stories

Recommended for you

Intensive behavior therapy no better than conventional support in treating teenagers with antisocial behavior

January 19, 2018
Research led by UCL has found that intensive and costly multisystemic therapy is no better than conventional therapy in treating teenagers with moderate to severe antisocial behaviour.

Babies' babbling betters brains, language

January 18, 2018
Babies are adept at getting what they need - including an education. New research shows that babies organize mothers' verbal responses, which promotes more effective language instruction, and infant babbling is the key.

College branding makes beer more salient to underage students

January 18, 2018
In recent years, major beer companies have tried to capitalize on the salience of students' university affiliations, unveiling marketing campaigns and products—such as "fan cans," store displays, and billboard ads—that ...

Inherited IQ can increase in early childhood

January 18, 2018
When it comes to intelligence, environment and education matter – more than we think.

Modulating molecules: Study shows oxytocin helps the brain to modulate social signals

January 17, 2018
Between sights, sounds, smells and other senses, the brain is flooded with stimuli on a moment-to-moment basis. How can it sort through the flood of information to decide what is important and what can be relegated to the ...

Baby brains help infants figure it out before they try it out

January 17, 2018
Babies often amaze their parents when they seemingly learn new skills overnight—how to walk, for example. But their brains were probably prepping for those tasks long before their first steps occurred, according to researchers.

1 comment

Adjust slider to filter visible comments by rank

Display comments: newest first

hush1
not rated yet Aug 16, 2011
Now target original origins - the brain activity and locality of the measured tasks were NOT in the beginning (of embryonic/fetal gestation developmental period) the origin sources of activity measured.

How to reduce and find the locality of original origins?
fMRI those born deaf and dumb.

Please sign in to add a comment. Registration is free, and takes less than a minute. Read more

Click here to reset your password.
Sign in to get notified via email when new comments are made.