How Do We Recognize Faces?

How do we recognize a face? Do we pick out “local” features— an eye or a mouth— and extrapolate from there? Or do we take in the “global” configuration—facial structure, distance between the features—at once?

Now, a group of psychologists— Sébastien Miellet and Philippe G. Schyns at the University of Glasgow, Scotland, and Roberto Caldara at the University of Fribourg in Switzerland—have settled the longstanding debate between scientists who hold to the “local” strategy and those who favor the “global” one.

“Face processing does not rely on a rigid system or a unique and mandatory information sampling strategy,” said Miellet. In fact, we use both the local and the global strategies to identify .

The findings will be published in an upcoming issue of , a journal of the Association for Psychological Science.

To collect their data, the researchers developed a method, called iHybrid, that pinpoints what information is used to identify a face. The technological aspects of the experiment were complex. In essence—and necessarily simplified—the researchers created hybrid images of pairs of famous men and manipulated them to create a face that looked, for instance, “like the natural child of Brad Pitt and William Macy,” as Miellet put it.

While an “eye-tracker” located the position of the subject’s gaze, iHybrid smoothly embedded the two faces in order to display one face around that gaze location and the other face in the surrounding area. The participants were asked what they saw. Because the image was a hybrid, they might name one or the other of the men, or neither.

The researchers gleaned a great deal of data, Miellet said. For one, the identity strategy followed from the initial “fixation point.” When people fixated first on the eye or mouth, they identified the face with a local strategy, even if their gaze subsequently moved. If the first fixation was in the center of the face, they were adopting global processing. But the same participant might use a global strategy in one trial and a local one in the next.

What did these data tell the researchers? Said Miellet: “Depending on viewing conditions — the first fixation location, the ambient light, the viewing angle or distance from the face — the system will adapt and do the most with the available information.” Like so much of the ways we see, learn, remember, and express ourselves, the process by which we recognize the people we know “is flexible.”

Related Stories

The kids are alright

May 26, 2011

Children should be seen and not heard... who says? A Philosophy academic at The University of Nottingham is challenging the adage by teaching primary school children to argue properly.

One year of the moon in 2.5 minutes

Jun 15, 2011

We don’t always have the time or ability to see the Moon every night of the year, but this video, from the Goddard Space Flight Center Scientific Visualization Studio, uses data from the Lunar Reconnaissance Orbiter ...

Recommended for you

Mother-daughter research team studies severe-weather phobia

Sep 19, 2014

No one likes severe weather, but for some just the thought of a thunderstorm, tornado, hurricane or blizzard can severely affect their lives. When blood pressures spike, individuals obsessively monitor weather forecasts and ...

Study: Pupil size shows reliability of decisions

Sep 18, 2014

Te precision with which people make decisions can be predicted by measuring pupil size before they are presented with any information about the decision, according to a new study published in PLOS Computational Bi ...

User comments

Adjust slider to filter visible comments by rank

Display comments: newest first

dutchman
not rated yet Jun 22, 2011
I though this was a gender related phenomenon. I read a long time go that men view people in general globally, where women view (scan) people's features. The intensity of the latter depended in the level of interest of the woman in the other person.

Wrong?