Are you one of those people who never forgets a face? New research from Vanderbilt University suggests that we can remember more faces than other objects and that faces "stick" the best in our short-term memory. The reason may be that our expertise in remembering faces allows us to package them better for memory.
"Our results show that we can store more faces than other objects in our visual short-term memory," Gauthier, associate professor of psychology and the study's co-author, says. "We believe this happens because of the special way in which faces are encoded."
Kim Curby, the study's primary author and a post-doctoral researcher at Yale University, likens such encoding to packing a suitcase. "How much you can fit in a bag depends on how well you pack it," she says. "In the same way, our expertise in 'packaging' faces means that we can remember more of them."
The findings, part of Curby's dissertation at Vanderbilt, are currently in press at the journal Psychonomic Bulletin and Review.
Curby and Gauthier's research has practical implications for the way we use visual short-term memory or VSTM. "Being able to store more faces in VSTM may be very useful in complex social situations," Gauthier says.
"This opens up the possibility of training people to develop similarly superior VSTM for other categories of objects," Curby adds.
Short-term memory is crucial to our impression of a continuous world, serving as temporary storage for information that we are currently using. For example, in order to understand this sentence, your short-term memory will remember the words in the beginning while you read through to the end. VSTM is a component of short-term memory that helps us process and briefly remember images and objects, rather than words and sounds.
VSTM allows us to remember objects for a few seconds, but its capacity is limited. Curby's and Gauthier's new research focuses on whether we can store more faces than other objects in VSTM, and the possible mechanisms underlying this advantage.
Participants studied up to five faces on a screen for varying lengths of time (up to four seconds). A single face was later presented and participants decided if this was a face that was part of the original display. For a comparison, the process was repeated with other objects, like watches or cars.
Curby and Gauthier found that when participants studied the displays for only a brief amount of time (half a second), they could store fewer faces than objects in VSTM. They believe this is because faces are more complex than watches or cars and require more time to be encoded. Surprisingly, when participants were given more time to encode the images (four seconds), an advantage for faces over objects emerged.
The researchers believe that our experience with faces explains this advantage. This theory is supported by the fact that the advantage was only obtained for faces encoded in the upright orientation, with which we are most familiar. Faces that were encoded upside-down showed no advantage over other objects.
"Our work is the first to show an advantage in capacity for faces over other objects," Gauthier explained. "Our results suggest that because experience leads you to encode upright faces in a different manner (not only using the parts, but using the whole configuration) you can store more faces in VSTM."
"What's striking about this is that some of the most prominent, current theories suggest that the capacity of VSTM is set in stone, unalterable by experience," Curby said. "However, our results clearly show that expert learning impacts VSTM capacity."
Curby and Gauthier plan to continue their research on VSTM processes. Their next step will focus on comparing VSTM capacity in people who are experts for other categories of complex objects, such as cars. Later, they will utilize brain imaging to pinpoint the mechanisms in the brain by which faces are encoded more efficiently than other objects.
Gauthier is a member of the Vanderbilt Vision Research Center, the Center for Integrative and Cognitive Neuroscience and the Vanderbilt Kennedy Center for Research on Human Development.
Source: Vanderbilt University