Researchers get humans to think like computers

ai
Credit: CC0 Public Domain

Computers, like those that power self-driving cars, can be tricked into mistaking random scribbles for trains, fences and even school busses. People aren't supposed to be able to see how those images trip up computers but in a new study, Johns Hopkins University researchers show most people actually can.

The findings suggest modern computers may not be as different from humans as we think, and demonstrate how advances in continue to narrow the gap between the visual abilities of people and machines. The research appears today in the journal Nature Communications.

"Most of the time, research in our field is about getting computers to think like people," says senior author Chaz Firestone, an assistant professor in Johns Hopkins' Department of Psychological and Brain Sciences. "Our project does the opposite—we're asking whether people can think like computers."

What's easy for humans is often hard for computers. Artificial intelligence systems have long been better than people at doing math or remembering large quantities of information; but for decades humans have had the edge at recognizing everyday objects such as dogs, cats, tables or chairs. But recently, "" that mimic the brain have approached the human ability to identify objects, leading to supporting self-driving cars, facial recognition programs and helping physicians to spot abnormalities in radiological scans.

But even with these technological advances, there's a critical blind spot: It's possible to purposely make images that neural networks cannot correctly see. And these images, called "adversarial" or "fooling" images, are a big problem: Not only could they be exploited by hackers and causes security risks, but they suggest that humans and machines are actually seeing images very differently.

Can you think like a machine? Try for yourself with this video game. Credit: Johns Hopkins University

In some cases, all it takes for a computer to call an apple a car, is reconfiguring a pixel or two. In other cases, machines see armadillos and bagels in what looks like meaningless television static.

"These machines seem to be misidentifying objects in ways humans never would," Firestone says. "But surprisingly, nobody has really tested this. How do we know people can't see what the computers did?"

To test this, Firestone and lead author Zhenglong Zhou, a Johns Hopkins senior majoring in cognitive science, essentially asked people to "think like a machine". Machines have only a relatively small vocabulary for naming images. So, Firestone and Zhou showed people dozens of fooling images that had already tricked computers, and gave people the same kinds of labeling options that the machine had. In particular, they asked people which of two options the computer decided the object was—one being the computer's real conclusion and the other a random answer. (Was the blob pictured a bagel or a pinwheel?) It turns out, people strongly agreed with the conclusions of the computers.

People chose the same answer as computers 75 percent of the time. Perhaps even more remarkably, 98 percent of people tended to answer like the computers did.

Next, researchers upped the ante by giving people a choice between the computer's favorite answer and its next-best guess. (Was the blob pictured a bagel or a pretzel?) People again validated the computer's choices, with 91 percent of those tested agreeing with the machine's first choice.

Even when the researchers had people guess between 48 choices for what the object was, and even when the pictures resembled television static, an overwhelming proportion of the subjects chose what the machine chose well above the rates for random chance. A total of 1,800 subjects were tested throughout the various experiments.

"We found if you put a person in the same circumstance as a computer, suddenly the humans tend to agree with the machines," Firestone says. "This is still a problem for artificial intelligence, but it's not like the is saying something completely unlike what a human would say."


Explore further

Research identifies key weakness in modern computer vision systems

Journal information: Nature Communications

Citation: Researchers get humans to think like computers (2019, March 22) retrieved 20 April 2019 from https://medicalxpress.com/news/2019-03-humans.html
This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no part may be reproduced without the written permission. The content is provided for information purposes only.
267 shares

Feedback to editors

User comments

Mar 22, 2019
In case you want to read the article (open access)
Humans can decipher adversarial images
Nature Communications volume 10, Article number: 1334 (2019)
https://www.natur...-08931-6

Mar 22, 2019
"We found if you put a person in the same circumstance as a computer, suddenly the humans tend to agree with the machines,"


But that wasn't the case at all. They put people in a situation where they had to guess what a computer would guess, which is a different problem than what the computer has to solve.

And on another point, the adversarial images cause computers to fail in bizarre ways that humans never would, like classifying a tortoise as a handgun. They fail in ways that just don't make any sense. The reverse test gives the game away by providing those false answers, and when people guess that the computer got it wrong, they pick the obviously false answer.

For the pictures that were just static noise, the question becomes about identifying some geometric feature, and matching it to a shape. If it's more like a figure eight, it's a pretzel, if it's more round, it's a bagel. The trick is, these answers are given. People might see something else.

Mar 22, 2019
Point being, the computer can only choose between a bagel and a pretzel, whereas a person might say "You know, that looks kinda like the olympic rings"

So when you do the reverse test, the person can only choose between bagel and pretzel, so they're forced to act like the computer, when in reality they wouldn't.

Please sign in to add a comment. Registration is free, and takes less than a minute. Read more