What you hear could depend on what your hands are doing

October 14, 2012

New research links motor skills and perception, specifically as it relates to a second finding—a new understanding of what the left and right brain hemispheres "hear." Georgetown University Medical Center researchers say these findings may eventually point to strategies to help stroke patients recover their language abilities, and to improve speech recognition in children with dyslexia.

The study, presented at Neuroscience 2012, the annual meeting of the Society for Neuroscience, is the first to match with left brain/right brain auditory processing tasks. Before this research, neuroimaging tests had hinted at differences in such processing.

"Language is processed mainly in the left hemisphere, and some have suggested that this is because the left hemisphere specializes in analyzing very rapidly changing sounds," says the study's senior investigator, Peter E. Turkeltaub, M.D., Ph.D., a in the Center for and Recovery. This newly created center is a joint program of Georgetown University and MedStar National Rehabilitation Network.

Turkeltaub and his team hid rapidly and slowly changing sounds in and asked 24 volunteers to simply indicate whether they heard the sounds by pressing a button.

"We asked the subjects to respond to sounds hidden in background noise," Turkeltaub explained. "Each subject was told to use their right hand to respond during the first 20 sounds, then their left hand for the next 20 second, then right, then left, and so on." He says when a subject was using their right hand, they heard the rapidly changing sounds more often than when they used their left hand, and vice versa for the slowly changing sounds.

"Since the left hemisphere controls the right hand and vice versa, these results demonstrate that the two hemispheres specialize in different kinds of sounds—the left hemisphere likes rapidly changing sounds, such as , and the likes slowly changing sounds, such as syllables or intonation," Turkeltaub explains. "These results also demonstrate the interaction between motor systems and perception. It's really pretty amazing. Imagine you're waving an American flag while listening to one of the presidential candidates. The speech will actually sound slightly different to you depending on whether the flag is in your left hand or your right hand."

Ultimately, Turkeltaub hopes that understanding the basic organization of auditory systems and how they interact with motor systems will help explain why language resides in the left hemisphere of the brain, and will lead to new treatments for language disorders, like aphasia (language difficulties after stroke or brain injury) or dyslexia.

"If we can understand the basic brain organization for audition, this might ultimately lead to new treatments for people who have speech recognition problems due to stroke or other brain injury. Understanding better the specific roles of the two hemispheres in auditory processing will be a big step in that direction. If we find that people with aphasia, who typically have injuries to the , have difficulty recognizing speech because of problems with low-level auditory perception of rapidly changing sounds, maybe training the specific auditory processing deficits will improve their ability to recognize speech," Turkeltaub concludes.

Explore further: Listen up: Abnormality in auditory processing underlies dyslexia

Related Stories

Recommended for you

Rat brain atlas provides MR images for stereotaxic surgery

October 21, 2016

Boris Odintsov, senior research scientist at the Biomedical Imaging Center at the Beckman Institute for Advanced Science and Technology at the University of Illinois in Urbana-Champaign, and Thomas Brozoski, research professor ...

ALS study reveals role of RNA-binding proteins

October 20, 2016

Although only 10 percent of amyotrophic lateral sclerosis (ALS) cases are hereditary, a significant number of them are caused by mutations that affect proteins that bind RNA, a type of genetic material. University of California ...

Imaging technique maps serotonin activity in living brains

October 20, 2016

Serotonin is a neurotransmitter that's partly responsible for feelings of happiness and for mood regulation in humans. This makes it a common target for antidepressants, which block serotonin from being reabsorbed by neurons ...

Overcoming egocentricity increases self-control

October 19, 2016

Neurobiological models of self-control usually focus on brain mechanisms involved in impulse control and emotion regulation. Recent research at the University of Zurich shows that the mechanism for overcoming egocentricity ...

1 comment

Adjust slider to filter visible comments by rank

Display comments: newest first

not rated yet Oct 14, 2012
Yes. I play piano. I switch (cross over my arms)to play the same piece of music with the hands now performing the original music with the left hand playing the right hand's role and vice versa.

You are correct. Basic brain organization for audition starts with conception. Fetal/embryonic motor/audition development get their 'cues' for development long before these innate functions mature to serve as the functions they are intended to perform at birth.

You will experience a delay in your understanding if lateralization of brain function is paramount to your understanding.

Please sign in to add a comment. Registration is free, and takes less than a minute. Read more

Click here to reset your password.
Sign in to get notified via email when new comments are made.