Wikipedia Bans AI-Generated Content


Wikipedia Bans AI-Generated Content

After months of heated debate and previous attempts to restrict the use of large language models on Wikipedia, on March 20 volunteer editors accepted a new policy that prohibits using them to create articles for the online encyclopedia. 

“Text generated by large language models (LLMs) often violates several of Wikipedia’s core content policies,” Wikipedia’s new policy states. “For this reason, the use of LLMs to generate or rewrite article content is prohibited, save for the exceptions given below.”

The new policy, which was accepted in an overwhelming 40 to 2 vote among editors, allows editors to use LLMs to suggest basic copyedits to their own writing, which can be incorporated into the article or rewritten after human review if the LLM doesn’t generate entirely new content on its own. 

“Caution is required, because LLMs can go beyond what you ask of them and change the meaning of the text such that it is not supported by the sources cited,” the policy states. “The use of LLMs to translate articles from another language’s Wikipedia into the English Wikipedia must follow the guidance laid out at Wikipedia:LLM-assisted translation.” 

I previously reported about editors using LLMs to translate Wikipedia articles and introducing errors to those articles in the process. 

Wikipedia editor, Ilyas Lebleu, who goes by Chaotic Enby on Wikipedia and who proposed the guideline said that it seemed unlikely the policy will last because previously the editor community has been divided on the issue. However, Lebleu said “The mood was shifting, with holdouts of cautious optimism turning to genuine worry.”

“A few months ago, a much more bare-bones guideline had passed, only banning the creation of brand new articles with LLMs,” Lebleu told me in an email. “A follow-up proposal to reword it into something more substantial failed to pass, but was noted to have ‘consensus for better guidelines along the lines of and/or in the spirit of this draft.’ In recent months, more and more administrative reports centered on LLM-related issues, and editors were being overwhelmed.”

The policy was written with the help of WikiProject AI Cleanup, a group of Wikipedia editors dedicated to finding and removing AI-generated errors on the site. Editors have been dealing with an increasing number of AI-generated articles or edits lately, and have made some minor adjustments to its guidelines as a result, like streamlining the process for removing AI-generated articles. Editors’ position, as well as the position of the Wikimedia Foundation, has been to not make blanket rules against AI because Wikipedia already uses some forms of automation, and because AI tools could assist editors in the future.  

The new policy doesn’t ban the use of other automated tools that are already in use or future implementations, but it does show the Wikipedia community is less optimistic about the benefit of AI-generated content, and taking a stand against it. 

“In context, this has implications far beyond Wikipedia,” Lebleu said. “The same flood of AI-generated content has been seen from social media to open-source projects, where agents submit pull requests much faster than human reviewers can keep up with. StackOverflow and the German Wikipedia paved the way in recent months with similar policies, and, as anxiety over the AI bubble grows, I foresee a domino effect,  empowering communities on other platforms to decide whether AI should be welcome. On their own terms.”

Leave a Reply

Your email address will not be published. Required fields are marked *