[Home]  [Headlines]  [Latest Articles]  [Latest Comments]  [Post]  [Sign-in]  [Mail]  [Setup]  [Help] 

Status: Not Logged In; Sign In

Would China Ever Invade Russia? Examining a Possible Scenario

Why Putin Can NEVER Use a Nuclear Weapon

Logical Consequence of Freedom4um point of view

Tucker Carlson: This current White House is being run by Satan, not human beings

U.S. Submarines Are Getting a Nuclear Cruise Missile Strike Capability: Destroyers Likely to Follow

Anti-Gun Cat Lady ATTACKS Congress Over Mexico & The UN!

Trump's new border czar will prioritize finding 300,000 missing migrant children who could be trafficking victims

Morgan Stanley: "If Musk Is Successful In Streamlining Government, It Would Broaden Earnings Growth And Stock Performance"

Bombshell Fauci Documentary Nails The Whole COVID Charade

TRUTH About John McCain's Service - Forgotten History

Bombshell Fauci Documentary Nails The Whole COVID Charade

Joe Rogan expressed deep concern that Joe Biden and Ukrainian President Zelensky will start World War III

Fury in Memphis after attempted murder suspect who ambushed FedEx employee walks free without bail

Tehran preparing for attack against Israel: Ayatollah Khamenei's aide

Huge shortage plagues Israeli army as losses mount in Lebanon, Gaza

Researchers Find Unknown Chemical In Drinking Water Posing "Potential Human Health Concern"

Putin visibly ‘shocked’ by US green-light for long-range missiles to strike inside Russia

The Problem of the Bitcoin Billionaires

Biden: “We’re leaving America in a better place today than when we came into office four years ago … "

Candace Owens: Gaetz out, Bondi in. There's more to this than you think.

OMG!!! Could Jill Biden Be Any MORE Embarrassing??? - Anyone NOTICE This???

Sudden death COVID vaccine paper published, then censored, by The Lancet now republished with peer review

Russian children returned from Syria

Donald Trump Indirectly Exposes the Jewish Neocons Behind Joe Biden's Nuclear War

Key European NATO Bases in Reach of Russia's Oreshnik Hypersonic Missile

Supervolcano Alert in Europe: Phlegraean Fields Activity Sparks Scientists Attention (Mass Starvation)

France reacted to the words of a US senator on sanctions against allies

Trump nominates former Soros executive for Treasury chief

SCOTUS asked to review if Illinois can keep counting mail-in ballots 2 weeks after election day

The Real Reason Government Workers Are Panicking About ElonÂ’s New Tracking System


Science/Tech
See other Science/Tech Articles

Title: Researchers Get Humans to Think Like Computers
Source: [None]
URL Source: [None]
Published: Mar 26, 2019
Author: staff
Post Date: 2019-03-26 23:58:19 by Tatarewicz
Keywords: None
Views: 62

TEHRAN (FNA)- Computers, like those that power self-driving cars, can be tricked into mistaking random scribbles for trains, fences and even school busses. People aren't supposed to be able to see how those images trip up computers but in a new study, Johns Hopkins University researchers show most people actually can.

The findings suggest modern computers may not be as different from humans as we think, and demonstrate how advances in artificial intelligence continue to narrow the gap between the visual abilities of people and machines. The research appears today in the journal Nature Communications.

"Most of the time, research in our field is about getting computers to think like people," says senior author Chaz Firestone, an assistant professor in Johns Hopkins' Department of Psychological and Brain Sciences. "Our project does the opposite -- we're asking whether people can think like computers."

What's easy for humans is often hard for computers. Artificial intelligence systems have long been better than people at doing math or remembering large quantities of information; but for decades humans have had the edge at recognizing everyday objects such as dogs, cats, tables or chairs. But recently, "neural networks" that mimic the brain have approached the human ability to identify objects, leading to technological advances supporting self-driving cars, facial recognition programs and helping physicians to spot abnormalities in radiological scans.

But even with these technological advances, there's a critical blind spot: It's possible to purposely make images that neural networks cannot correctly see. And these images, called "adversarial" or "fooling" images, are a big problem: Not only could they be exploited by hackers and causes security risks, but they suggest that humans and machines are actually seeing images very differently.

In some cases, all it takes for a computer to call an apple a car, is reconfiguring a pixel or two. In other cases, machines see armadillos and bagels in what looks like meaningless television static.

"These machines seem to be misidentifying objects in ways humans never would," Firestone says. "But surprisingly, nobody has really tested this. How do we know people can't see what the computers did?"

To test this, Firestone and lead author Zhenglong Zhou, a Johns Hopkins senior majoring in cognitive science, essentially asked people to "think like a machine." Machines have only a relatively small vocabulary for naming images. So, Firestone and Zhou showed people dozens of fooling images that had already tricked computers, and gave people the same kinds of labeling options that the machine had. In particular, they asked people which of two options the computer decided the object was -- one being the computer's real conclusion and the other a random answer. (Was the blob pictured a bagel or a pinwheel?) It turns out, people strongly agreed with the conclusions of the computers.

People chose the same answer as computers 75 percent of the time. Perhaps even more remarkably, 98 percent of people tended to answer like the computers did.

Next, researchers upped the ante by giving people a choice between the computer's favorite answer and its next-best guess. (Was the blob pictured a bagel or a pretzel?) People again validated the computer's choices, with 91 percent of those tested agreeing with the machine's first choice.

Even when the researchers had people guess between 48 choices for what the object was, and even when the pictures resembled television static, an overwhelming proportion of the subjects chose what the machine chose well above the rates for random chance. A total of 1,800 subjects were tested throughout the various experiments.

"We found if you put a person in the same circumstance as a computer, suddenly the humans tend to agree with the machines," Firestone says. "This is still a problem for artificial intelligence, but it's not like the computer is saying something completely unlike what a human would say."

Post Comment   Private Reply   Ignore Thread  



[Home]  [Headlines]  [Latest Articles]  [Latest Comments]  [Post]  [Sign-in]  [Mail]  [Setup]  [Help]