When AI Steers Us Astray

A new debugging tool can pinpoint errors that cause neural networks to make mistakes

Join Our Community of Science Lovers!

As artificial-intelligence systems become widespread, the chances grow that their glitches will have dangerous consequences. For example, scientists at the Massachusetts Institute of Technology recently fooled a Google-trained AI program into identifying a plastic toy turtle as a rifle. If a future robot cop or soldier made such an error, the results could be tragic. But researchers are now developing tools to sniff out potential flaws among the billions of virtual “brain cells” that make up such systems.

Many image-recognition programs, car autopilots and other forms of AI use artificial neural networks, in which components dubbed “neurons” are fed data and cooperate to solve a problem—such as spotting obstacles in the road. The network “learns” by repeatedly adjusting the connections between its neurons and trying the problem again. Over time, the system determines which neural connection patterns are best at computing solutions. It then adopts these as defaults, mimicking how the human brain learns.

A key challenge of this technology is that developers often do not know how networks arrive at their decisions. This can make it hard to figure out what went wrong when a mistake is made, says Junfeng Yang, a computer scientist at Columbia University and a co-author of a new study presented last October at a symposium in Shanghai.


On supporting science journalism

If you're enjoying this article, consider supporting our award-winning journalism by subscribing. By purchasing a subscription you are helping to ensure the future of impactful stories about the discoveries and ideas shaping our world today.


Yang and his colleagues created DeepXplore, a program designed to debug AI systems by reverse engineering their learning processes. It tests a neural network with a wide range of confusing real-world inputs and tells the network when its responses are wrong so that it can correct itself. For example, DeepXplore could determine if a camera image fed to a car-driving AI system mistakenly steered a vehicle toward pedestrians. The debugging tool also monitors which neurons in a network are active and tests each one individually. Previous AI-debugging tools could not tell if every neuron had been checked for errors, Yang says.

In tests on 15 state-of-the-art neural networks—including some used in self-driving cars and computer malware detection—DeepXplore discovered thousands of bugs missed by earlier technology. It boosted the AI systems' overall accuracy by 1 to 3 percent on average, bringing some systems up to 99 percent. DeepXplore's technique could help developers build “more accurate and more reliable” neural networks, says Shan Lu, a computer scientist at the University of Chicago, who did not take part in the new research. This approach, Lu adds, “in turn can benefit many scientific research disciplines and our daily lives.”

Charles Q. Choi is a frequent contributor to Scientific American. His work has also appeared in The New York Times, Science, Nature, Wired, and LiveScience, among others. In his spare time, he has traveled to all seven continents.

More by Charles Q. Choi
Scientific American Magazine Vol 318 Issue 3This article was published with the title “'I'm Sorry, Dave'” in Scientific American Magazine Vol. 318 No. 3 (), p. 20
doi:10.1038/scientificamerican0318-20b

It’s Time to Stand Up for Science

If you enjoyed this article, I’d like to ask for your support. Scientific American has served as an advocate for science and industry for 180 years, and right now may be the most critical moment in that two-century history.

I’ve been a Scientific American subscriber since I was 12 years old, and it helped shape the way I look at the world. SciAm always educates and delights me, and inspires a sense of awe for our vast, beautiful universe. I hope it does that for you, too.

If you subscribe to Scientific American, you help ensure that our coverage is centered on meaningful research and discovery; that we have the resources to report on the decisions that threaten labs across the U.S.; and that we support both budding and working scientists at a time when the value of science itself too often goes unrecognized.

In return, you get essential news, captivating podcasts, brilliant infographics, can't-miss newsletters, must-watch videos, challenging games, and the science world's best writing and reporting. You can even gift someone a subscription.

There has never been a more important time for us to stand up and show why science matters. I hope you’ll support us in that mission.

Thank you,

David M. Ewalt, Editor in Chief, Scientific American

Subscribe