Researchers surprised to find how neural circuits identify information needed for decisions

November 6, 2013, Stanford University
After studying biological data and a computer model, Stanford scientists believe that neurons in the prefrontal cortex receive both color and motion data, and screen out the irrelevant sensory input to make decisions. This confounds the conventional thinking that such decisions involved one group of neurons that filtered out irrelevant signals and a second group of neurons that made a decision based on the relevant data. This graphic shows how one group of neurons in the prefrontal cortex "multitask" to make decisions. Color and motion signals converge in this area. But when experimenters asked a question involving color, a selection vector directed color signals to a line attractor. This line attractor represents the neurons 'concentrating' on the relevant signal and disregarding irrelevant motion data. A split second later these same multitasking neurons made the color-based decision. Credit: David Sussillo, Shenoy Lab, Stanford Engineering

While eating lunch you notice an insect buzzing around your plate. Its color and motion could both influence how you respond. If the insect was yellow and black you might decide it was a bee and move away. Conversely, you might simply be annoyed at the buzzing motion and shoo the insect away. You perceive both color and motion, and decide based on the circumstances. Our brains make such contextual decisions in a heartbeat. The mystery is how.

In an article published November 7th in the journal Nature, a team of Stanford neuroscientists and engineers delve into this -making process and report some findings that confound the conventional wisdom.

Until now, neuroscientists have believed that decisions of this sort involved two steps: one group of neurons that performed a gating function to ascertain whether motion or color was most relevant to the situation, and a second group of neurons that considered only the sensory input relevant to making a decision under the circumstances.

But in a study that combined brain recordings from trained monkeys, and a sophisticated computer model based on that biological data, Stanford neuroscientist William Newsome and three coauthors discovered that the entire decision-making process may occur in a localized region of the prefrontal cortex.

In this region of the brain, located in the frontal lobes just behind the forehead, they found that color and motion signals converged in a specific circuit of neurons. Based on their experimental evidence and computer simulations, the scientists hypothesized that these neurons act together to make two snap judgments: is color or motion the most relevant sensory input in the current context, and what action to take as a result.

"We were quite surprised," said Newsome, the Harman Family Provostial Professor at the Stanford School of Medicine and lead author.

He and first author Valerio Mante, a former Stanford neurobiologist now at the University of Zurich and the Swiss Federal Institute of Technology, had begun the experiment expecting to find that the irrelevant signal, whether color or motion, would be gated out of the circuit long before the decision-making neurons went into action.

"What we saw instead was this complicated mix of signals that we could measure, but whose meaning and underlying mechanism we couldn't understand," Newsome said. "These signals held information about the color and motion of the stimulus, which stimulus dimension was most relevant, and the decision that the monkeys made. But the signals were profoundly mixed up at the single neuron level. We decided there was a lot more we needed to learn about these neurons, and that the key to unlocking the secret might lie in a population level analysis of the circuit activity."

To solve this brain puzzle the neurobiologists began a cross-disciplinary collaboration with Krishna Shenoy, a professor of electrical engineering at Stanford, and David Sussillo, co-first author on the paper and a postdoctoral scholar in Shenoy's lab.

Sussillo created a software model to simulate how these neurons worked. The idea was to build a model sophisticated enough to mimic the decision-making process, but easier to study than taking repeated electrical readings from a brain.

The general model architecture they used is called a recurrent neural network: a set of software modules designed to accept inputs and perform tasks similar to how biological neurons operate. The scientists designed this artificial neural network using computational techniques that enabled the software model to make itself more proficient at decision-making over time.

"We challenged the artificial system to solve a problem analogous to the one given to the monkeys," Sussillo explained. "But we didn't tell the neural network how to solve the problem."

As a result, once the artificial network learned to solve the task, the scientists could study the model to develop inferences about how the biological neurons might be working.

The entire process was grounded in the biological experiments.

The neuroscientists trained two macaque monkeys to view a random-dot visual display that had two different features—motion and color. For any given presentation, the dots could move to the right or left, and the color could be red or green. The monkeys were taught to use sideways glances to answer two different questions depending on the currently instructed "rule" or context. Were there more red or green dots (ignore the motion)? Or, were the dots moving to the left or right (ignore the color)?

Eye-tracking instruments recorded the glances, or saccades, that the monkeys used to register their responses. Their answers were correlated with recordings of neuronal activity taken directly from an area in the prefrontal cortex known to control saccadic eye movements.

The neuroscientists collected 1,402 such experimental measurements. Each time the monkeys were asked one or the other question. The idea was to obtain brain recordings at the moment when the monkeys saw a visual cue that established the context (either the red/green or left/right question), and what decision the animal made regarding color or direction of motion.

It was the puzzling mish-mash of signals in the brain recordings from these experiments that prompted the scientists to build the recurrent neural network as a way to rerun the experiment, in a simulated way, time and time again.

As the four researchers became confident that their software simulations accurately mirrored the actual biological behavior, they studied the model to learn exactly how it solved the task. This allowed them to form a hypothesis about what was occurring in that patch of neurons in the prefrontal cortex where perception and decision occurred.

"The idea is really very simple," Sussillo explained.

Their hypothesis revolves around two mathematical concepts: a line attractor and a selection vector.

The entire group of neurons being studied received sensory data about both the color and the motion of the dots.

The line attractor is a mathematical representation for the amount of information that this group of neurons was getting about either of the relevant inputs, color or motion.

The selection vector represented how the model responded when the experimenters flashed one of the two questions: red or green, left or right?

What the model showed was that when the question pertained to color, the selection vector directed the artificial neurons to accept color information while ignoring the irrelevant motion information. Color data became the line attractor. After a split second these neurons registered a decision, choosing the red or green answer based on the data they were supplied.

If question was about motion, the selection vector directed motion information to the line attractor and the artificial neurons chose left or right.

"The amazing part is that a single neuronal circuit is doing all of this," Sussillo says. "If our model is correct, then almost all neurons in this biological circuit appear to be contributing to almost all parts of the information selection and decision-making mechanism."

Newsome put it like this: "We think that all of these are interested in everything that's going on, but they're interested to different degrees. They're multitasking like crazy."

Other researchers who are aware of the work but were not directly involved are commenting on the paper.

"This is a spectacular example of excellent experimentation combined with clever data analysis and creative theoretical modeling," said Larry Abbott, Co-Director of the Center for Theoretical Neuroscience and the William Bloor Professor, Neuroscience, Physiology & Cellular Biophysics, Biological Sciences at Columbia University.

Christopher Harvey, a professor of neurobiology at Harvard Medical School, said the paper "provides major new hypotheses about the inner-workings of the prefrontal cortex, which is a brain area that has frequently been identified as significant for higher cognitive processes but whose mechanistic functioning has remained mysterious."

The Stanford scientists are now designing a new biological experiment to ascertain whether the interplay between selection vector and line attractor, which they deduced from their software model, can be measured in actual brain signals.

"The model predicts a very specific type of neural activity under very specific circumstances," Sussillo said. "If we can stimulate the prefrontal cortex in the right way, and then measure this activity, we will have gone a long way to proving that the model mechanism is indeed what is happening in the biological circuit."

The four researchers worked together on this paper for more than 18 months at the James H. Clark Center, home of Bio-X, Stanford's ongoing effort to stimulate interdisciplinary collaborations in the field of human biology.

Explore further: NYU's Movshon winner of 'Golden Brain' award for research on the neuroscience of vision

More information: Context-dependent computation by recurrent dynamics in prefrontal cortex, Nature, 2013.

Related Stories

NYU's Movshon winner of 'Golden Brain' award for research on the neuroscience of vision

November 6, 2013
New York University Professor J. Anthony Movshon has been named the recipient of Minerva Foundation's 2013 Golden Brain Award "for his foundational contributions to the field of visual neuroscience," the Berkeley, Calif.-based ...

Neurons that can multitask greatly enhance the brain's computational power, study finds

May 20, 2013
Over the past few decades, neuroscientists have made much progress in mapping the brain by deciphering the functions of individual neurons that perform very specific tasks, such as recognizing the location or color of an ...

Antidepressant drug induces a juvenile-like state in neurons of the prefrontal cortex

November 4, 2013
For long, brain development and maturation has been thought to be a one-way process, in which plasticity diminishes with age. The possibility that the adult brain can revert to a younger state and regain plasticity has not ...

Neural balls and strikes: Where categories live in the brain

January 15, 2012
Hundreds of times during a baseball game, the home plate umpire must instantaneously categorize a fast-moving pitch as a ball or a strike. In new research from the University of Chicago, scientists have pinpointed an area ...

Neuroscientists show how decision-making processes are influenced by neurons

January 15, 2013
Whether in society or nature, decisions are often the result of complex interactions between many factors. Because of this it is usually difficult to determine how much weight the different factors have in making a final ...

Recommended for you

Even without nudging blood pressure up, high-salt diet hobbles the brain

January 16, 2018
A high-salt diet may spell trouble for the brain—and for mental performance—even if it doesn't push blood pressure into dangerous territory, new research has found.

Brain imaging predicts language learning in deaf children

January 15, 2018
In a new international collaborative study between The Chinese University of Hong Kong and Ann & Robert H. Lurie Children's Hospital of Chicago, researchers created a machine learning algorithm that uses brain scans to predict ...

Preterm babies may suffer setbacks in auditory brain development, speech

January 15, 2018
Preterm babies born early in the third trimester of pregnancy are likely to experience delays in the development of the auditory cortex, a brain region essential to hearing and understanding sound, a new study reveals. Such ...

BOLD view of white matter

January 15, 2018
The brain consists of gray matter, which contains the nerve cell bodies (neurons), and white matter, bundles of long nerve fibers (axons) that until recently were considered passive transmitters of signals between different ...

Does an exploding brain network cause chronic pain?

January 12, 2018
A new study finds that patients with fibromyalgia have brain networks primed for rapid, global responses to minor changes. This abnormal hypersensitivity, called explosive synchronization (ES), can be seen in other network ...

An innovative PET tracer can measure damage from multiple sclerosis in mouse models

January 12, 2018
The loss or damage of myelin, a cellular sheath that surrounds and insulates nerves, is the hallmark of the immune-mediated neurological disorder multiple sclerosis (MS). When segments of this protective membrane are damaged, ...


Adjust slider to filter visible comments by rank

Display comments: newest first

Parag Jasani
1 / 5 (2) Nov 07, 2013
What makes us capable of making decisions? As per the Dichotomized Operating System of model of mind, we make decisions using Cumulative Feedback Mechanism (CFM). CFM is a mechanism designed to make the most out of complex operations offered by human life through judgment based processing, which is one of its main functions. It is named so because it makes decisions in a cumulative manner using a feedback mechanism. More at http://www.whatis...FMI.aspx
not rated yet Nov 07, 2013
Interesting research.

But lay off the ads for personal ideas, this is a science site.
1 / 5 (1) Nov 11, 2013
"The model predicts a very specific type of neural activity under very specific circumstances," Sussillo said.

And this model shows how this arises:

Senses of location to organize color and motion of the processes associated with the input of color and motion and the storage (memory) of color/motion input.

The "line vector" is a single sheath or module (see illustration from the link) at the 'receiving end' or summation of all other sheaths or modules the researchers here label "selection vector".

Resolution of input determines the "selection vector". Obviously, an event or object below the threshold of resolution of any sensory input also determines a "selection vector".

(You may not see the motions or see the colors of an event or object like a wasp. The sound of an event /object like a wasp is a "selection vector" already summing additional "selection vectors" (from memory) to the "line vector".
Go Stanford

Please sign in to add a comment. Registration is free, and takes less than a minute. Read more

Click here to reset your password.
Sign in to get notified via email when new comments are made.