How Do We Recognize Faces?

June 21, 2011

How do we recognize a face? Do we pick out “local” features— an eye or a mouth— and extrapolate from there? Or do we take in the “global” configuration—facial structure, distance between the features—at once?

Now, a group of psychologists— Sébastien Miellet and Philippe G. Schyns at the University of Glasgow, Scotland, and Roberto Caldara at the University of Fribourg in Switzerland—have settled the longstanding debate between scientists who hold to the “local” strategy and those who favor the “global” one.

“Face processing does not rely on a rigid system or a unique and mandatory information sampling strategy,” said Miellet. In fact, we use both the local and the global strategies to identify .

The findings will be published in an upcoming issue of , a journal of the Association for Psychological Science.

To collect their data, the researchers developed a method, called iHybrid, that pinpoints what information is used to identify a face. The technological aspects of the experiment were complex. In essence—and necessarily simplified—the researchers created hybrid images of pairs of famous men and manipulated them to create a face that looked, for instance, “like the natural child of Brad Pitt and William Macy,” as Miellet put it.

While an “eye-tracker” located the position of the subject’s gaze, iHybrid smoothly embedded the two faces in order to display one face around that gaze location and the other face in the surrounding area. The participants were asked what they saw. Because the image was a hybrid, they might name one or the other of the men, or neither.

The researchers gleaned a great deal of data, Miellet said. For one, the identity strategy followed from the initial “fixation point.” When people fixated first on the eye or mouth, they identified the face with a local strategy, even if their gaze subsequently moved. If the first fixation was in the center of the face, they were adopting global processing. But the same participant might use a global strategy in one trial and a local one in the next.

What did these data tell the researchers? Said Miellet: “Depending on viewing conditions — the first fixation location, the ambient light, the viewing angle or distance from the face — the system will adapt and do the most with the available information.” Like so much of the ways we see, learn, remember, and express ourselves, the process by which we recognize the people we know “is flexible.”

Related Stories

Recommended for you

Neural efficiency hypothesis confirmed

July 27, 2015

One of the big questions intelligence researchers grapple with is just how differences in intelligence are reflected in the human brain. Researchers at ETH Zurich have succeeded in studying further details relating to suspected ...

Your phone knows if you're depressed

July 15, 2015

You can fake a smile, but your phone knows the truth. Depression can be detected from your smartphone sensor data by tracking the number of minutes you use the phone and your daily geographical locations, reports a small ...

1 comment

Adjust slider to filter visible comments by rank

Display comments: newest first

dutchman
not rated yet Jun 22, 2011
I though this was a gender related phenomenon. I read a long time go that men view people in general globally, where women view (scan) people's features. The intensity of the latter depended in the level of interest of the woman in the other person.

Wrong?

Please sign in to add a comment. Registration is free, and takes less than a minute. Read more

Click here to reset your password.
Sign in to get notified via email when new comments are made.