A More Reliable Wikipedia Could Come from AI Research Assistants

A neural network can identify Wikipedia references that are unlikely to support an article’s claims—and scour the Web for better sources

··· Person holding smartphone with logo of online encyclopedia Wikipedia on screen in front of website. Focus on phone display.

AI tools could save time for editors checking the accuracy of Wikipedia entries.

Join Our Community of Science Lovers!

Wikipedia lives and dies by its references, the links to sources that back up information in the online encyclopaedia. But sometimes, those references are flawed — pointing to broken websites, erroneous information or non-reputable sources.

A study published on 19 October in Nature Machine Intelligence suggests that artificial intelligence (AI) can help to clean up inaccurate or incomplete reference lists in Wikipedia entries, improving their quality and reliability.

Fabio Petroni at London-based company Samaya AI and his colleagues developed a neural-network-powered system called SIDE, which analyses whether Wikipedia references support the claims they’re associated with, and suggests better alternatives for those that don’t.


On supporting science journalism

If you're enjoying this article, consider supporting our award-winning journalism by subscribing. By purchasing a subscription you are helping to ensure the future of impactful stories about the discoveries and ideas shaping our world today.


“It might seem ironic to use AI to help with citations, given how ChatGPT notoriously botches and hallucinates citations. But it’s important to remember that there’s a lot more to AI language models than chatbots,” says Noah Giansiracusa, who studies AI at Bentley University in Waltham, Massachusetts.

AI filter

SIDE is trained to recognize good references using existing featured Wikipedia articles, which are promoted on the site and receive a lot of attention from editors and moderators.

It is then able to identify claims within pages that have poor-quality references through its verification system. It can also scan the Internet for reputable sources, and rank options to replace bad citations.

To put the system to the test, Petroni and his colleagues used SIDE to suggest references for featured Wikipedia articles that it had not seen before. In nearly 50% of cases, SIDE’s top choice for a reference was already cited in the article. For the others, it found alternative references.

When SIDE’s results were shown to a group of Wikipedia users, 21% preferred the citations found by the AI, 10% preferred the existing citations and 39% did not have a preference.

The tool could save time for editors and moderators checking the accuracy of Wikipedia entries, but only if it is deployed correctly, says Aleksandra Urman, a computational communication scientist at the University of Zurich, Switzerland. “The system could be useful in flagging those potentially-not-fitting citations,” she says. “But then again, the question really is what the Wikipedia community would find the most useful.”

Urman points out that the Wikipedia users who tested the SIDE system were twice as likely to prefer neither of the references as they were to prefer the AI-suggested ones. “This would mean that in these cases, they would still go and search for the relevant citation online,” she says.

This article is reproduced with permission and was first published on October 19, 2023.

It’s Time to Stand Up for Science

If you enjoyed this article, I’d like to ask for your support. Scientific American has served as an advocate for science and industry for 180 years, and right now may be the most critical moment in that two-century history.

I’ve been a Scientific American subscriber since I was 12 years old, and it helped shape the way I look at the world. SciAm always educates and delights me, and inspires a sense of awe for our vast, beautiful universe. I hope it does that for you, too.

If you subscribe to Scientific American, you help ensure that our coverage is centered on meaningful research and discovery; that we have the resources to report on the decisions that threaten labs across the U.S.; and that we support both budding and working scientists at a time when the value of science itself too often goes unrecognized.

In return, you get essential news, captivating podcasts, brilliant infographics, can't-miss newsletters, must-watch videos, challenging games, and the science world's best writing and reporting. You can even gift someone a subscription.

There has never been a more important time for us to stand up and show why science matters. I hope you’ll support us in that mission.

Thank you,

David M. Ewalt, Editor in Chief, Scientific American

Subscribe