This article has been reviewed according to Science X's editorial process and policies. Editors have highlighted the following attributes while ensuring the content's credibility:


peer-reviewed publication

trusted source


Researchers help unravel brain processes involved in vision

Credit: CC0 Public Domain

Faced with images that break the expected pattern, like a do not enter sign where a stop sign is expected, how does the brain react and learn compared to being shown images that match what was predicted?

That was the question a team, including York University, set out to answer. A long-standing theory suggests the learns a predictive model of the world, and its internal predictions are updated when incoming sensory data proves them wrong. However, what the researchers found surprised them, says York Faculty of Science Associate Professor Joel Zylberberg, co-corresponding author of the newly published paper.

"Testing this theory has always been a challenge," he says. "We needed to be able to measure the top-down signals to the sensory areas of the brain over long periods of time to show how the brain learns new sensory input patterns."

Using a , the researchers displayed images of visual patterns over multiple days and then presented other images that violated those patterns while measuring the brain's activity in the visual cortex, where from the retina is processed. The idea was to test how the neurons reacted to the new pattern-violating sensory information.

Several of the researchers, including Zylberberg, are Fellows in the Canadian Institute for Advanced Research's Learning in Machines and Brains group, which conducted the research as part of the Allen Institute for Brain Science's Brain Observatory and its OpenScope program. OpenScope has been compared to an observatory where astronomers work together to study the universe; only this time, researchers are sharing data to study the brain.

The measurements were taken at the neurons' distal apical dendrites of the , which receive top-down signals, and at their cell bodies, which receive bottom-up signals. They wanted to know if the distal apical dendrites processed differently from their cell bodies when the signals both matched and violated expected patterns.

It turns out the brain's response to image patterns that violate the brain's predictions evolves differently over time when compared to pattern-matching images.

"Surprisingly, the distal apical dendrites responses grew significantly over time, becoming increasingly sensitive to inputs that violate the patterns, while the cell bodies lost their initially strong sensitivity," says Zylberberg, a computational neuroscientist. "This finding could offer critical insight into sensory computation and predictive learning in the brain."

The finding suggests that the pattern-violating stimuli drove the changes, and different forms of pattern-violating stimuli may elicit different kinds of prediction errors than expected. It points to a component of the brain that could have a distinct and important role in sensory learning not previously known.

"Knowing how the brain processes new visual is important for developing better machine learning algorithms and applications which could hopefully help restore people's sight in the future," says Zylberberg.

The paper was published in JNeurosci

More information: Colleen Gillen et al, Responses to pattern-violating visual stimuli evolve differently over days in somata and distal apical dendrites, JNeurosci (2023). DOI: 10.1523/JNEUROSCI.1009-23.2023

Journal information: Journal of Neuroscience
Provided by York University
Citation: Researchers help unravel brain processes involved in vision (2023, November 21) retrieved 4 March 2024 from
This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no part may be reproduced without the written permission. The content is provided for information purposes only.

Explore further

Dendrites may help neurons perform complicated calculations


Feedback to editors