Researchers publish study on the flexibility of sensory perception

Researchers publish study on the flexibility of sensory perception
Professor Dr. Christoph Kayser. Credit: Universitaet Bielefeld

Hearing, sight, touch – our brain captures a wide range of distinct sensory stimuli and links them together. The brain has a kind of built-in filter function for this: sensory impressions are only integrated if it is necessary and useful for the task at hand. Researchers from Bielefeld University, Oxford University (Great Britain), and Aix-Marseille University (France) investigated this phenomenon of flexibility in perception, and have now published a study on their findings that appears in the scientific journal Neuron. In their publication, the researchers reveal where sensory stimuli are integrated in the brain, and in which area of the brain this flexibility can be located. From Bielefeld University, Professor Dr. Christoph Kayser and Dr. Hame Park from the Cluster of Excellence CITEC were involved in the study.

"We are interested in how the sensory input," says Kayser, who heads the "Cognitive Neuroscience" research group. In his work, Kayser deals with multi-sensory integration – the combination of various sensory data. This happens, for instance, when watching a movie: you hear what the characters are saying to each other while at the same time watching the movements of their lips. It is not always useful, however, for auditory and to be automatically integrated in the : one example of this would be watching a foreign-language film that is dubbed and the movements of the actors' lips do not match the spoken sounds.

The researchers' study was looking to identify the areas of the brain in which flexible sensory integration takes place. For this, they tested out three potential models. While different sensory were processed completely apart from each other in the first , they were automatically integrated in the second model. The third variant was the model of "causal inference" in which different sensory stimuli are only integrated if they are not distant from each other in spatial or chronological terms. For example, if you always hear a sound and see an image at the same time, the brain integrates this . However, if the sound and image appear together, they will not be integrated – even though they were previously separate from each other. "In the 'causal inference' model, the brain infers that the source of the sensory stimuli might be the same. Sensory stimuli are therefore not automatically integrated – this is only the case if they do originate from the same source," says Kayser.

To test these three models, were exposed to visual and auditory stimuli. In the experiment, light and sound stimuli were shown, at times simultaneously, and at other times with different frequencies. Meanwhile, the researchers recorded the brain activity of their study participants using magnetoencephalography (MEG). They found that the three models match up to different areas in the brain, and thus correspond with distinct levels of processing. At the lowest level, sensory information is mapped separately in the visual and auditory cortexes. Following this, this information is automatically integrated in the parietal lobe, which is located in the upper area of the brain. Only at a higher level of processing does the brain parse out the information from the previous stages and, if necessary, filter out disruptive sensory stimuli. This flexibility in perception is located in special areas of the frontal lobe that are responsible for abstract thinking. "It has long been known how people deal with different sensory information – but at the level of behavior. With our study, we were able to show for the first time how and where in the brain this kind of information is processed," says Kayser.

The findings of the study could be of use in many other areas of research, particularly in work on abstract thinking, as flexibility and causal relationships play an important role in this process. "How the brain processes is relevant for technical applications, including the integration of humans and machines," says Kayser. This is a key topic for Kayser's colleagues at Bielefeld University's Cluster of Excellence CITEC. The findings of this study are also important in the clinical context, where the findings could be used to help better understand diseases in which patients have difficulty in correctly processing sensory information, such as autism.


Explore further

Multitasking amygdala neurons respond to sights, sounds, and touch

More information: Yinan Cao et al. Causal Inference in the Multisensory Brain, Neuron (2019). DOI: 10.1016/j.neuron.2019.03.043
Journal information: Neuron

Citation: Researchers publish study on the flexibility of sensory perception (2019, May 1) retrieved 18 July 2019 from https://medicalxpress.com/news/2019-05-publish-flexibility-sensory-perception.html
This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no part may be reproduced without the written permission. The content is provided for information purposes only.
7 shares

Feedback to editors

User comments

Please sign in to add a comment. Registration is free, and takes less than a minute. Read more