We’re now sadly used to seeing toxic exchanges play out on social media platforms like X (formerly Twitter), Facebook and TikTok.
But Wikipedia is a reference work. How heated can people get over an encyclopedia?
Our research, published today, shows the answer is very heated. For example, one Wikipedia editor wrote to another:
i will find u in real life and slit your throat.
That’s a problem for many reasons, but chief among them is if Wikipedia goes down in a ball of toxic fire, it might take the rest of the internet’s information infrastructure with it.
The internet’s favourite encyclopedia
In some ways, Wikipedia is both an encyclopedia and a social media platform.
It’s the fourth most popular website on the internet, behind only such giants as Google, YouTube and Facebook.
Every day, millions of people worldwide use it for quick fact-checks or in-depth research.
And what happens to Wikipedia matters beyond the platform itself because of its central role in online information infrastructure.
Google search relies heavily on Wikipedia and the quality of its search results would decrease substantially if Wikipedia disappeared.
But it’s not just an increasingly authoritative source of knowledge. Even though we don’t always lump Wikipedia in with other social media platforms, it shares some common features.
Wikipedia is edited by volunteers from all over the world and hosted by a nonprofit organization, the @Wikimedia Foundation, that supports editors and ensures Wikipedia remains secure, fast, and free for anyone to access and use. (2/4) pic.twitter.com/eNSVhGRqwo
— Wikipedia (@Wikipedia) November 15, 2023
It relies on contributors to create the content that the public will view and it creates spaces for those contributors to interact. Wikipedia relies solely on the work of volunteers: no one is paid for writing or editing content.
Moreover, no one checks the credentials of editors — anyone can make a contribution. This arguably makes Wikipedia the most successful collaborative project in history.
However, the fact that Wikipedia is a collaborative platform also makes it vulnerable.
A 2015 survey found 38% of surveyed Wikipedia users had experienced harassment on the platform.
What if the collaborative environment deteriorates, and its volunteer editors abandon the project?
What effect do toxic comments have on Wikipedia’s editors, content and community?
Abusive comments lead to disengaging
To answer this question, we started with Wikipedia’s “user’s talk pages”. A user’s talk page is a place where other editors can interact with the user. They can post messages, discuss personal topics, or extend discussions from an article’s talk page.
Every editor has a personal user’s talk page, and the majority of toxic comments made on the platform are on these pages.
We collected information on 57 million comments made on the user’s talk pages of 8.5 million editors across the six most active language editions of Wikipedia (English, German, French, Italian, Spanish and Russian) over a period of 20 years.
We then used a state-of-the-art machine learning algorithm to identify toxic comments. The algorithm looked for attributes a human might consider toxic, like insults, threats, or identity attacks.
We compared the activity of editors before and after they received a toxic comment, as well as with a control group of similar editors who received a non-toxic rather than toxic comment.
We found receiving a single toxic comment could reduce an editor’s activity by 1.2 active days in the short term. Considering that 80,307 users on English Wikipedia alone have received at least one toxic comment, the cumulative impact could amount to 284 lost human-years.
Moreover, some users don’t just contribute less. They stop contributing altogether.
We found that the probability of leaving Wikipedia’s community of contributors increases after receiving a toxic comment, with new users being particularly vulnerable. New editors who receive toxic comments are nearly twice as likely to leave Wikipedia as would be expected otherwise.
Wikipedia is just as vulnerable to toxic commentary as other popular websites. Wikimedia Commons, CC BY-SA
Wide-ranging consequences
This matters more than you might think to the millions who use Wikipedia.
First, toxicity likely leads to poorer-quality content on the site. Having a diverse editor cohort is a crucial factor for maintaining content quality. The vast majority of Wikipedian editors are men, which is reflected in the content on the platform.
There are fewer articles about women, which are shorter than articles about men and more likely to centre on romantic relationships and family-related issues. They are also more often linked to articles about the opposite gender. Women are often described as wives of famous people rather than for their own merits, for example.
While multiple barriers confront women editors on Wikipedia, toxicity is likely one of the key factors that contributes to the gender imbalance. Although men and women are equally likely to face online harassment and abuse, women experience more severe violations and are more likely to be affected by such incidents, including self-censoring.
This may affect other groups as well: our research showed that toxic comments often include not just gendered language but also ethnic slurs and other biases.
Finally, a significant rise in toxicity, especially targeted attacks on new users, could jeopardise Wikipedia’s survival.
Following a period of exponential growth in its editor base during the early 2000s, the number has been largely stable since 2016, with the exception of a brief activity spike during the COVID pandemic. Currently about the same number of editors join the project as leave, but the balance could be easily tipped if the people left because of online abuse.
That would damage not only Wikipedia, but also the rest of the online information infrastructure it helps to support.
There’s no easy fix to this, but our research shows promoting healthy communication practices is critical to protecting crucial online information ecosystems.


Microsoft Azure Growth Forecast Beats Expectations Amid Rising AI Competition
Amazon Stock Rises as Meta Expands AWS Partnership for AI Infrastructure
DeepSeek Slashes AI Model Pricing to Boost Adoption and Challenge Global Rivals
Meta Raises 2026 Capex Outlook Amid AI Spending Surge, Shares Drop After Earnings
Advantest Stock Falls on Weak Outlook Despite Strong AI-Driven Results
Australia Targets Meta, Google, and TikTok With New News Payment Tax Proposal
$16B Michigan Data Center Project Boosts U.S. AI Infrastructure Expansion
Taiwan Court Fines Tokyo Electron Unit $4.78M in Major TSMC Trade Secrets Case
U.S. Demand for Alternative Satellite Providers Remains Strong Amid SpaceX Regulatory Push
Seagate Stock Surges After Strong Q3 Earnings Beat and Bullish Outlook
Qualcomm Stock Surges Despite Weak Guidance After Q2 2026 Earnings Beat
Judge Dismisses Elon Musk’s Fraud Claims Against OpenAI, Trial to Proceed on Remaining Allegations
T-Mobile Beats Q1 Earnings Expectations on Strong Postpaid Growth
Intel Stock Surges as AI Chip Demand Drives Strong Q2 Forecast
Taiwan Activates Backup Communications After Undersea Cable Break on Dongyin Island
Samsung Reports Record Profit as AI Boom Drives Memory Chip Demand 




