This is a humorousessay. It containshumorous advice or opinions of one or more Wikipedia contributors and isn't meant to be taken seriously. This is not an encyclopedia article or one ofWikipedia's policies or guidelines and may not representcommunity consensus. |
| This essay, entitled "Keep Wikipedia Human: The Case Against LLM-Generated Articles," was written by an AI and was pasted into this page by a Wikipedia editor and then edited for review before submission. |
In the age oflarge language models (LLMs) likeChatGPT, it might be tempting to use these tools to rapidly generate Wikipedia articles. Their fluent prose and speedy output present an alluring shortcut. However, the Wikipedia editing community is overwhelminglyopposed to letting LLMs write our content. The reason is simple: such AI-generated text threatens the core principles that make Wikipedia reliable. We have seen that asking an LLM to “write a Wikipedia article” can produceoutright fabrication, complete withfictitious references. From factual inaccuracies that readers might never catch to citations that lead nowhere, LLM-created content poses risks Wikipedia cannot afford. This essay outlines the community’s firm concerns –factual inaccuracy andhallucinations, unverifiable sourcing and fake citations, damage to the collaborative editing model, and long-term erosion of content integrity – and argues that Wikipedia must remain a human-driven project grounded in verifiable truth.
One of the most alarming issues with AI-generated text is the tendency of LLMs to “hallucinate” – to fabricate information that sounds plausible but is entirely false. By design, an LLM predicts words based on patterns, not on an understanding of facts.[1] This means an LLM will sometimes produce content thatreads convincingly but has no basis in reality. It might even invent details when faced with a prompt on an obscure topic or a nonsensical request. For example, in one trial an LLM confidently described the habits of a completelyfictitious animal as if it were real, blending real-sounding facts into a coherent but false narrative.[2] Theentire statement was false, yet it was delivered authoritatively – a clear illustration of how AI can offerconvincing lies with a confident tone.[3]
Such hallucinated content isn’t just a theoretical edge case; it’s a practical nightmare for Wikipedia. Even if90% of an AI-generated article is correct and 10% is false, that 10% is a huge problem in an encyclopedia. Wikipedia’s reputation hinges on accuracy. One subtly false claim or distorted quote can misinform readers and damage trust. LLMs have no sense of responsibility or context – they’ll cheerfully generate an “article” about the health benefits of eating crushed glass if asked. They might also incorporate hidden biases or defamatory assertions without realizing it. In short, hallucinations from LLMs introduce outright falsehoods into what should be a repository of verified knowledge. Wikipedia editors view this as an unacceptable risk. The community has long held thatverifiability, not truth (in other words, verifiable evidence over unsupported claims) is non-negotiable. Content that “sounds true” is worthless to Wikipedia if it can’t be verified or if it’s simply concocted by an algorithm.
Wikipedia’s bedrock principle ofverifiability requires that material be backed byreliable, published sources that readers can check. Every fact on Wikipedia should be attributable; if it isn’t attributable, it doesn’t belong. Here lies a fundamental conflict with LLM-generated content: LLMs often donot follow this principle. They might output paragraphs with no citations at all, or citeunreliable sources, or even citesources that don’t exist.[4] Indeed, in many observed cases, LLMs have producedacademic-sounding references that are completely fabricated.[5] A human editor might spend hours searching for a cited journal article or book, only to discover it was never real. Wikipedia’s verifiability policy is clear that readers“must be able to check that any of the information... is not just made up” – but with AI hallucinations, the informationis made up, and no real source exists to support it.
This problem has already manifested in experiments. Wikipedia editors who have tested LLMs report that these tools tend to sprinkle in references thatlook plausible but are bogus. In one analysis, researchers found an AI-written draft of a historical article included seven references – andfive of those were fabricated.[5] The citationsappeared credible, naming reputable journals and authors, but on closer inspection they led nowhere. Such fake citations are more insidious than an obvious “citation needed” tag; they give a false impression ofcredibility, undermining the reader’s ability to trust Wikipedia. Even outside Wikipedia, hallucinated citations have caused scandals – for instance, legal briefs and academic papers generated by AI have cited cases and articles that don’t exist.[5] For Wikipedia, this is a verifiability crisis: content backed by phantom sources isworse than unsourced content, because it actively misleads. It violates the core policy that all challenged materialmust be supported by inline citations to reliable sources. If an AI tool cannot guarantee genuine, checkable sources, it has no place writing our articles. The community’s stance is that every sentence added to Wikipedia needs humanverification against real,reliable sources – a standard an LLM cannot meet on its own.[6]

Beyond the immediate content issues, using LLMs to generate articles poses a serious threat to Wikipedia’s collaborative editing ecosystem. The site thrives on acommunity of volunteers who write,fact-check, and refine articles throughconsensus anddebate. Introducing masses of machine-generated text upends this model in several ways.
First, it creates an overwhelmingmaintenance burden on human editors. Wikipedia’s volunteers have limited time and energy. If they must sift through AI-produced drafts line-by-line to weed out errors, or chase down phantom references, it turns editing into a chore of cleanup rather than a process of building knowledge. As one community essay notes, Wikipedia operates on an informal social contract: editors put significant effort into their contributions so that others“do not need to clean up after them”. LLM-generated content shatters that contract. Poorly vetted AI textincreases the maintenance burden on other volunteers, effectively asking humans to become janitors for machine output. This is not why people volunteer for Wikipedia, and it’s not sustainable. Indeed, editors have reported feeling“flooded non-stop with horrendous drafts” created by AI, full of mistakes that require significant time to fix. One experienced editor described many AI-written drafts as containing“lies and fake references” that demand painstaking correction.[7] This tsunami of low-quality content isn’t just annoying – it threatens tooverwhelm the quality control processes that keep Wikipedia reliable.
Second, over-reliance on LLM content coulderode the culture of collaboration. Wikipedia is built by people discussing and reasoning together. An AI cannot participate in talk page debates, cannot understand nuance, and cannot exercise judgment about neutrality or due weight of a topic. If editors start deferring to AI-generated text, the role of human deliberation diminishes. There is also a risk of mistrust and social disruption: editors may grow suspicious of each other’s contributions (“Was this written by a bot?”), makinggood-faith collaboration harder. Already, the community has felt the need to developguides for spotting AI-generated writing and even a dedicatedWikiProject AI Cleanup to coordinate the identification and removal of AI content.[7] While these efforts demonstrate commendable vigilance, they also reveal a worrying shift – volunteer time that used to go towards improving content is now spent onpolicing AI-generated additions. Seasoned editors now have to train newcomers on how to recognize the telltale signs of “AI slop” (unnatural phrasing, excessive formality, oddly formatted citations, etc.)[8] rather than on how to research and write well.[9][7] This is anopportunity cost: every hour spent fighting AI-generated nonsense is an hour not spent writing a featured article or reviewing actual sourced content.
Thehuman cost of this fight is real. Longtime contributors report that patrolling for AI fakery has made their work less enjoyable and more exhausting. What used to be a rewarding hobby of knowledge-sharing is at risk of becoming, in the words of some, an “increasingly investigative and adversarial process.”[9] The community fears volunteer burnout as editors struggle to keep up with a flood of machine-produced material. If contributing ceases to be fun or fulfilling, many will simplywalk away. Many already have.[10] Wikipedia has always depended on the passion and dedication of its volunteers – if that passion is sapped by endless cleanup and suspicion, the entire project’s health is at stake.
Finally, allowing AI-written articles coulddeter the development of human expertise within the community.[3] Traditionally, new editors learn by researching sources, writing text, and getting feedback from peers. If instead a newbie simply clicks a button to have an AI “write” an article, they bypass the crucial learning process and likely introduce errors. When their contributions are then deleted or heavily corrected, they may become discouraged and leave, rather than growing into proficient Wikipedians. In this sense, LLMs present a false promise of easy contribution that can actually alienate good-faith newcomers when the community (rightly) rejects the subpar AI content. In the long run, we risk losing a generation of editors if we normalize automating the very tasks that build editorial skills and community bonds.LLMs are tools, not replacements for human judgment,[1] and relying on them too heavily in article-writing undermines the very human collaboration that built Wikipedia. The editing community is stronglyunited on this point: Wikipedia should not become a dumping ground for unchecked machine output, because doing so would hollow out the values and social framework that sustain the project.
The abstract risks of AI-generated content become starkly clear when looking at concrete examples. In recent years, Wikipedia editors have uncovered multiple cases of AI-written content slipping into the site – with problematic results. Below are a few illustrative episodes that highlight how LLM-generated text can fail disastrously:

These cases and thought-experiments underscore why the community has responded so aggressively against AI-generated content. The “fortress” hoax was deleted under speedy deletion criteria as soon as it was identified, and it prompted calls for tighter controls. In fact, by 2025 English Wikipedia updated its deletion policy to allowimmediate removal of LLM-generated pages that lack human review (criterionG15). Similarly, the detection of AI-drafted biased articles has led to more rigorous new article patrols and scrutiny of sources. When the Wikimedia Foundation trialed an AI tool called "Simple Article Summaries" that would do as the name suggests, the community’s reaction was swift and negative – editors called it a“ghastly idea” that would erode trust, given AI’s tendency to err and hallucinate.[7][17] OneArs Technica report quoted an editor’s visceral reaction: “Yuck,” capturing the general sentiment.[18] The trial was paused almost immediately amid this backlash.[7][19] The message from these examples is crystal clear: when it comes to creating content, Wikipedia’s volunteer editors have learned through hard experience thatautomation can fail spectacularly. Each failure, whether real or hypothetical, reinforces the lesson that without human-level understanding and rigorous source-checking, Wikipedia articles generated by AI areticking time bombs for our reputation.
After examining the issues – hallucinated facts, unverifiable citations, disruption to collaboration, and concrete examples of AI-generated failures – the conclusion is firm:Wikipedia must prioritize human collaboration and verifiability over the temptations of automation. The integrity of this project, built over two decades, relies on information being accurate and sources being real. That integrity cannot be maintained if we open the floodgates to machine-generated text that nobody has fully vetted. In the end, an article’s worth on Wikipedia is not how quickly it was produced, but how true and reliable it is. LLMs offer speed, not judgment; volume, not discernment. They do not have a reputation to uphold –we do.
The Wikipedia editing community’s stance is not born oftechnophobia, but of experience and dedication to our mission. We have seen the dangers firsthand. We know that once trust is broken – if readers begin to doubt that Wikipedia’s content is thoroughly checked and sourced – the damage is long-term. As one editor put it during the AI summary debate, deploying unvetted AI content would do“immediate and irreversible harm to our readers and to our reputation as a decently trustworthy and serious source.” Wikipedia has become synonymous with a certain level of reliability (“sober boringness,” as that editor wryly noted, and that is a strength we must protect.
You generallyshouldn't use AI to edit Wikipedia. This essay, minus this section, was written to look like the output of an AI chatbot (and to be truthful, some parts of itwere AI-generated). One of the most obvious ways you can tell is by the tone, as well as the fact that some things arebolded when they shouldn't be. The references within were designed to look at first like they could be real, but in fact some of them are obviously fake when you take a look at them. ChatGPT has no concept of what counts as a reliable source, and it also can't easily tell the difference between truth and fiction. This applies to Bing, Gemini, and whatever other AI chatbots may be developed in the future. As forme, I've personally only really used it for translation ofcertainarticles written in alanguage I don't speak to English, or perhaps to help find sources for something, and that's basically the only really acceptable use. If you do go ahead and decide to use AI anyway, you do soat your own risk. Chances are that people will be able to tell, and you may even getblocked. You'll wind up creating a lot of work for us human editors to sift through to clean up after you. So do us all a favor.
{{cite book}}: CS1 maint: location missing publisher (link)