New brain-scanning technique allows scientists to see when and where the brain processes visual information

by Anne Trafton
MIT researchers combined fMRI and MEG data to reveal which parts of the brain are active shortly after an image is seen. At around 60 milliseconds, only early visual cortex in the back of the brain was active (image at left). Then, activity spread to brain regions involved in later visual processing until the inferior temporal cortex was activated (image at right). This brain region represents complex shapes and categories of objects. Credit: MIT

Every time you open your eyes, visual information flows into your brain, which interprets what you're seeing. Now, for the first time, MIT neuroscientists have noninvasively mapped this flow of information in the human brain with unique accuracy, using a novel brain-scanning technique.

This technique, which combines two existing technologies, allows researchers to identify precisely both the location and timing of human brain activity. Using this new approach, the MIT researchers scanned individuals' brains as they looked at different images and were able to pinpoint, to the millisecond, when the brain recognizes and categorizes an object, and where these processes occur.

"This method gives you a visualization of 'when' and 'where' at the same time. It's a window into processes happening at the millisecond and millimeter scale," says Aude Oliva, a principal research scientist in MIT's Computer Science and Artificial Intelligence Laboratory (CSAIL).

Oliva is the senior author of a paper describing the findings in the Jan. 26 issue of Nature Neuroscience. Lead author of the paper is CSAIL postdoc Radoslaw Cichy. Dimitrios Pantazis, a research scientist at MIT's McGovern Institute for Brain Research, is also an author of the paper.

When and where

Until now, scientists have been able to observe the location or timing of human brain activity at high resolution, but not both, because different imaging techniques are not easily combined. The most commonly used type of brain scan, functional magnetic resonance imaging (fMRI), measures changes in blood flow, revealing which parts of the brain are involved in a particular task. However, it works too slowly to keep up with the brain's millisecond-by-millisecond dynamics.

Another imaging technique, known as magnetoencephalography (MEG), uses an array of hundreds of sensors encircling the head to measure magnetic fields produced by neuronal activity in the brain. These sensors offer a dynamic portrait of brain activity over time, down to the millisecond, but do not tell the precise location of the signals.

To combine the time and location information generated by these two scanners, the researchers used a computational technique called representational similarity analysis, which relies on the fact that two similar objects (such as two human faces) that provoke similar signals in fMRI will also produce similar signals in MEG. This method has been used before to link fMRI with recordings of neuronal electrical activity in monkeys, but the MIT researchers are the first to use it to link fMRI and MEG data from human subjects.

In the study, the researchers scanned 16 human volunteers as they looked at a series of 92 images, including faces, animals, and natural and manmade objects. Each image was shown for half a second.

"We wanted to measure how flows through the brain. It's just pure automatic machinery that starts every time you open your eyes, and it's incredibly fast," Cichy says. "This is a very complex process, and we have not yet looked at higher cognitive processes that come later, such as recalling thoughts and memories when you are watching objects."

Each subject underwent the test multiple times—twice in an fMRI scanner and twice in an MEG scanner—giving the researchers a huge set of data on the timing and location of . All of the scanning was done at the Athinoula A. Martinos Imaging Center at the McGovern Institute.

Millisecond by millisecond

By analyzing this data, the researchers produced a timeline of the brain's object-recognition pathway that is very similar to results previously obtained by recording electrical signals in the visual cortex of monkeys, a technique that is extremely accurate but too invasive to use in humans.

About 50 milliseconds after subjects saw an image, visual information entered a part of the brain called the primary visual cortex, or V1, which recognizes basic elements of a shape, such as whether it is round or elongated. The information then flowed to the inferotemporal cortex, where the brain identified the object as early as 120 milliseconds. Within 160 milliseconds, all objects had been classified into categories such as plant or animal.

The MIT researchers are now using representational similarity analysis to study the accuracy of computer models of vision by comparing brain scan data with the models' predictions of how vision works.

Using this approach, scientists should also be able to study how the analyzes other types of information such as motor, verbal, or sensory signals, the researchers say. It could also shed light on processes that underlie conditions such as memory disorders or dyslexia, and could benefit patients suffering from paralysis or neurodegenerative diseases.

"This is the first time that MEG and fMRI have been connected in this way, giving us a unique perspective," Pantazis says. "We now have the tools to precisely map brain function both in space and time, opening up tremendous possibilities to study the human ."

More information: Resolving human object recognition in space and time, DOI: 10.1038/nn.3635

Related Stories

Training your brain using neurofeedback

Jan 21, 2014

A new brain-imaging technique enables people to 'watch' their own brain activity in real time and to control or adjust function in pre-determined brain regions. The study from the Montreal Neurological Institute ...

Mapping objects in the brain

Dec 20, 2013

A brain region that responds to a particular category of objects is found to consist of small clusters of neurons encoding visual features of these objects.

Recommended for you

New viral tools for mapping brains

53 minutes ago

(Medical Xpress)—A brain-computer-interphase that is optogenetically-enabled is one of the most fantastic technologies we might envision today. It is likely that its full power could only be realized under ...

Link seen between seizures and migraines in the brain

16 hours ago

Seizures and migraines have always been considered separate physiological events in the brain, but now a team of engineers and neuroscientists looking at the brain from a physics viewpoint discovered a link ...

Neuroscience: Why scratching makes you itch more

22 hours ago

Turns out your mom was right: Scratching an itch only makes it worse. New research from scientists at Washington University School of Medicine in St. Louis indicates that scratching causes the brain to release ...

User comments

Please sign in to add a comment. Registration is free, and takes less than a minute. Read more

Click here to reset your password.
Sign in to get notified via email when new comments are made.