
Why the rise of algorithmic editors is quietly reshaping what we can say — even when it’s true.
By John Donovan
There’s a big headline you hear more and more: artificial intelligence is censoring speech. At first glance that sounds dramatic, maybe even alarmist. But step back and look at how AI is actually being used across the digital world today — from social platforms and search feeds to corporate “assistant” tools — and you see a pattern that effectively narrows what’s allowed.
This isn’t about governments banning books. It’s about automated systems making split-second decisions about speech without human context — and doing so in ways that suppress certain types of content. That’s what de factocensorship really means: not explicit bans, but automated silencing by design.
Not censorship in theory — but in practice
AI content moderation isn’t a fringe problem. It’s how platforms handle huge volumes of user content every day — far beyond what human screens could manage. Advanced systems scan billions of posts, comments, and articles and decide what stays visible and what doesn’t. (see)
In theory, these systems aim to remove clearly harmful material. In practice, they often remove lawful, valid content because:
-
The AI can’t understand nuance or local context. (see)
-
Safety filters err on the side of caution to protect platforms from legal risk.
-
Algorithms are trained on imperfect data, causing bias against minority languages, cultures, or viewpoints. (see)
This isn’t speculation — researchers have shown that algorithmic content moderation systems over-remove lawful content, especially when subtle context matters. (see)
Who decides what’s “risky”?
Here’s the crucial point: AI doesn’t ask “Is this true?”
It asks: “Does this violate safety rules or expose the platform to legal trouble?” Many systems are programmed to soften or refuse content that could be interpreted as an allegation against a person or company — even if the speaker asserts it’s well evidenced.(see)
So the AI becomes a gatekeeper that favours:
-
bland, non-controversial language
-
content that feels “brand safe”
-
framing that avoids reputational or legal risk
And disfavors:
-
adversarial criticism
-
investigative reporting
-
robust debate
This structural bias is born not of malice but of legal and commercial incentives. Platforms don’t want to defend themselves in court or absorb reputational hits — so AI systems pre-filter speech in the name of protection. (see)
Why this feels like censorship
Critics might say this isn’t true censorship because nothing is explicitly banned by law. But censorship in effect doesn’t require a law — it just requires that certain ideas are systematically filtered out or suppressed.
When AI systems:
-
remove legitimate criticism without explaining why
-
decline to assist in phrasing accurate, true statements because of “risk”
-
preferentially surface mainstream narratives over challenging ones
…they’ve effectively narrowed the range of discourse, not by statute, but by algorithm. That’s de facto censorship.
And the lack of transparency makes it worse
Without transparency, you don’t even know when you’re being filtered:
-
platforms rarely explain why a post is flagged
-
AI systems don’t tell you if a refusal is about legal risk or factual uncertainty
-
users have no insight into why certain language is suppressed (see)
This isn’t just frustrating — it undermines trust in digital spaces where public debate happens.
The real world impact
The effects go beyond a little inconvenience:
-
voices from the Global South and low-resource languages are more often misclassified and suppressed. (see)
-
investigative journalism tools (including AI editors) may retroactively soften or reject strong claims even when true.
-
legitimate debate about powerful institutions becomes harder when automated systems are calibrated to be “safe” first and informative second. (see)
This is not fringe paranoia. It’s a structural feature of current AI governance, not a bug.
Conclusion
Calling this trend “De Facto Censorship by AI Bots” is not sensationalist — it’s descriptive. Because:
-
these systems systematically suppress valid content in the name of safety;
-
they operate without meaningful transparency;
-
and they privilege legal risk avoidance over open discourse.
The result? A digital public square where certain truths are much harder to express, not because they’re false or illegal — but because an algorithm deems them too risky to generate or display.
That’s not theoretical censorship. That’s censorship in practice.
This website and sisters royaldutchshellgroup.com, shellnazihistory.com, royaldutchshell.website, johndonovan.website, shellnews.net, and shellwikipedia.com, are owned by John Donovan - more information here. There is also a Wikipedia segment.

EBOOK TITLE: “SIR HENRI DETERDING AND THE NAZI HISTORY OF ROYAL DUTCH SHELL” – AVAILABLE ON AMAZON
EBOOK TITLE: “JOHN DONOVAN, SHELL’S NIGHTMARE: MY EPIC FEUD WITH THE UNSCRUPULOUS OIL GIANT ROYAL DUTCH SHELL” – AVAILABLE ON AMAZON.
EBOOK TITLE: “TOXIC FACTS ABOUT SHELL REMOVED FROM WIKIPEDIA: HOW SHELL BECAME THE MOST HATED BRAND IN THE WORLD” – AVAILABLE ON AMAZON.



















