Welcome

Welcome to Grappy's Soap Box - a platform for insightful commentary on politics, media, free speech, climate change, and more, focusing on Australia, the USA, and global perspectives.

Friday, 23 January 2026

Wikipedia, Bias, and the Poisoning of Our Future Knowledge




For years we were told that Wikipedia was the great democratiser of knowledge. A free, open encyclopedia written by volunteers, corrected by the wisdom of crowds, and guided by the noble principle of “neutral point of view.”

That comforting myth is now collapsing.

Two recent pieces — a detailed academic paper Toxic Truth: How Wikipedia Poisons Global Knowledge and an investigation by HonestReporting — lay bare what many of us have suspected for some time. Wikipedia is no longer a neutral reference work. It has become a highly politicised gatekeeper of “truth”, shaped by activist editors, ideological capture, and in some cases by foreign state interests.

And the danger is no longer confined to Wikipedia itself.

Today, Wikipedia is one of the primary training sources for large language models — the very AI systems that will soon answer our children’s questions, write our news summaries, and provide “authoritative” explanations on everything from history to medicine to geopolitics.

If Wikipedia is biased, then tomorrow’s AI will be biased — permanently, invisibly, and at scale.


Captured From Within

The Toxic Truth paper documents something deeply troubling. Wikipedia is not shaped by millions of casual contributors. It is controlled by a relatively small group of highly motivated editors who dominate sensitive political topics.

These editors decide which sources are “reliable,” which viewpoints are “fringe,” and which facts are “undue.” In theory this is meant to protect quality. In practice it allows ideological activists to quietly rewrite history.

On contentious subjects — Israel, terrorism, race, gender, climate, COVID — the same pattern appears. Critical voices are removed. Alternative perspectives are downgraded. Language is carefully adjusted to frame one side as legitimate and the other as suspect.

This is not censorship by force. It is something more effective: narrative control disguised as neutrality.


Foreign Influence and Qatar’s Shadow

The HonestReporting investigation takes this even further.

It reveals how Qatar — a regime that funds Hamas, hosts extremist clerics, and runs the propaganda network Al Jazeera — has built deep influence inside Wikipedia’s editorial ecosystem.

Editors linked to Qatari interests have shaped articles on the Middle East, terrorism, and Israel for years. Sources hostile to Israel are elevated. Israeli perspectives are minimised or framed as propaganda. Terror groups are softened into “militants” or “fighters.”

All while Wikipedia continues to present itself as a neutral educational charity.

This matters because Wikipedia is no longer just an encyclopedia. It is becoming the backbone of global digital knowledge.


When AI Learns From a Corrupted Source

Here is the truly alarming part.

Modern AI systems are trained on massive datasets, and Wikipedia is one of their core reference sources. Not one of many. One of the most important.

That means every distortion, every omission, every ideological framing embedded in Wikipedia today will be replicated tomorrow across countless AI platforms.

Future students may never read Wikipedia directly. But they will read AI summaries trained on Wikipedia.

If Wikipedia teaches that Israel is uniquely evil, that Western democracies are colonial oppressors, that Islamist violence is “resistance,” or that certain scientific debates are settled beyond discussion — then that becomes the default worldview of artificial intelligence.

And unlike Wikipedia, AI will not show its sources. You won’t know what has been filtered out.

Bias will become invisible.


The Greatest Danger: Monopoly on Truth

The real problem is not that Wikipedia contains errors. All reference works do.

The problem is that Wikipedia is rapidly becoming the single source of truth.

Schools rely on it. Journalists consult it. Search engines rank it at the top. AI systems ingest it wholesale.

When one platform becomes the foundation of knowledge, ideological capture becomes catastrophic.

A biased newspaper can be challenged by another newspaper.
A biased academic can be challenged by another academic.
But when the reference layer itself is compromised, the entire knowledge stack above it becomes distorted.

That is not just misinformation.

That is civilisational risk.


A Warning We Cannot Ignore

The authors of Toxic Truth and the investigators at HonestReporting are not arguing for censorship. They are arguing for transparency, pluralism, and accountability.

Wikipedia must not be allowed to present activism as neutrality.
AI developers must not be allowed to train on politically contaminated data without disclosure.
And governments, universities, and educators must stop treating Wikipedia as an unquestioned authority.

Because if we allow one ideologically captured platform to define reality for both humans and machines, we will not be living in an information age.

We will be living in an engineered one.

And once artificial intelligence learns a poisoned version of truth, correcting it later may prove impossible.

(I strongly recommend reading the attached paper. This is not an academic curiosity. It is about who controls knowledge itself. https://honestreporting.com/wikipedia-qatar-and-the-future-of-knowledge/ )

Wednesday, 21 January 2026

Trump's Economic Impact Over His First Year

Donald Trump commenced his second presidential term exactly one year ago. It has been a whirlwind, with ne'er a day without some controversy. Everyone has a view on Trump and will readily voice their support or opposition to one or more of his edicts. He has done more, much more, than any other President, certainly in recent history. Given this is a critical year with the mid-terms due in less than 11 months, we can look at his impact on the key drivers of November's vote.

I have gathered the following graphs from a recent article on The Epoch Times titled Trump's First Year by the Numbers

_____________________________________________________________











Tuesday, 20 January 2026

Adelaide Writers' Festival Cancels Itself

Once again, cancel culture has devoured one of its own — and this time, I’m not shedding a tear.

The Adelaide Writers’ Festival, long a comfortable home for fashionable left-wing causes and anti-Israel rhetoric, has imploded in spectacular fashion. What began as a belated attempt at moral clarity ended with mass walk-outs, the resignation of its director, and the cancellation of the entire event. It would be hard to script a better example of progressive self-destruction.

The controversy centred on Dr Randa Abdel-Fattah, a long-time anti-Israel activist who had been scheduled as a featured speaker. Following the Bondi massacre — in which 15 Jews were murdered in an Islamist terror attack — festival organisers quietly removed her from the program, citing “cultural sensitivity concerns.”

Those concerns were hardly imaginary.

As HonestReporting documented in a recent presentation, Abdel-Fattah had mocked Israelis fleeing the Nova music festival on October 7 while Hamas terrorists were still rampaging through southern Israel. The following day she made an image of Hamas paraglider terrorists her social-media cover photo. More recently, she was filmed teaching Australian schoolchildren to chant “Israel is a terrorist state” and “From the river to the sea.”

Let’s be clear: that isn’t education. It’s ideological indoctrination and the normalisation of hate.

These were more than sufficient reasons to conclude that she had no place on a public literary platform, particularly only weeks after Australian Jews were massacred on their own soil. Yet the moment she was removed, a predictable backlash erupted.

Other writers rushed to her defence. Zionists were blamed. “Cancel culture” was invoked. The removal of a Hamas apologist was framed as censorship rather than basic moral hygiene. Incredibly, the protesters portrayed her as the victim — not the Jewish community that had just buried its dead.

The protest escalated into a boycott by participating writers. Under mounting political pressure — including from South Australia’s Labor Premier — the festival director folded. The result? A mass walk-out, her resignation, and the cancellation of the entire festival.

And then came the final insult: the organisers rescinded their original statement and announced that Abdel-Fattah would be reinvited for next year’s festival.

You couldn’t make this up.

This was never about “cultural sensitivity.” It was about whether an institution funded by the public should platform someone who openly glorifies terrorists, mocks massacre victims, and teaches children to chant genocidal slogans. The real question is not why she was removed — it’s why the Adelaide Writers’ Festival ever thought she belonged there in the first place.

For years, the festival had no problem hosting anti-Israel speakers. It even “uninvited” a pro-Israel speaker in the past after pressure from activists. So when, for once, it showed the faintest flicker of moral awareness, its own ideological tribe turned on it.

The result is poetic justice.

Either you believe in open dialogue and pluralism — or you don’t. Either all opinions are allowed — or only the approved ones. The Adelaide Writers’ Festival tried to straddle both worlds. In the end, it chose none.

Good riddance.

If this is what passes for “literary culture” in 2026 Australia — censoring one side, platforming terror apologists, and collapsing into hysterics the moment minimal standards are applied — then perhaps the country is better off without it.

I will include the HonestReporting video below this post. It’s worth watching. It documents, in grim detail, how a supposedly enlightened cultural institution managed to disgrace itself in record time.

Sometimes cancel culture doesn’t just cancel speakers.

Sometimes it cancels itself.











Monday, 19 January 2026

Weekly Roundup - Top Articles and Commentary from Week 4 of 2026

 


Here are links to some selected articles of interest and our posts from this week.





We welcome all feedback; please feel free to submit your comments or contact me via email at grappysb@gmail.com or on X at @grappysb 

A Normal Day in 2040 — And How Robots Made It Better





Much of the public conversation about humanoid robots is framed in fear. Jobs will disappear. Humans will become obsolete. Society will unravel.

It’s a familiar pattern. We said the same about tractors, washing machines, personal computers and the internet. In every case, technology didn’t end work — it changed work. And it made everyday life better.

So instead of dystopian speculation, let’s imagine something far more radical: an ordinary day in the life of an ordinary family in 2040.


6:30 AM — A Calm Start to the Day

Mark and Sarah wake up to a quiet house. No frantic rushing. No piles of laundry. No sink full of dishes from the night before.

Their household robot — a standard, affordable humanoid assistant now as common as a dishwasher once was — has already done the overnight maintenance. It folded clothes, cleaned the kitchen, charged itself, restocked groceries ordered automatically the previous evening, and prepared breakfast.

Coffee is ready. Toast is warm. The kitchen is spotless.

Sarah skims the news while eating. Mark checks his schedule. Neither of them has lifted a finger yet — and that’s the point.


8:00 AM — Work, Still Human

Mark still works — just not in a factory or warehouse. He’s a project coordinator for a renewable energy company, managing teams, planning infrastructure upgrades, and solving problems that still require human judgment, creativity and accountability.

Sarah is a speech therapist, working with children who have learning difficulties. No robot can replace empathy, nuanced communication, or the trust built between a therapist and a child.

The robots didn’t eliminate meaningful work. They eliminated drudgery.

Both parents leave the house knowing it will remain clean, secure and running smoothly all day without human effort.


9:00 AM — Education, Transformed

Their two children, Emma (12) and Leo (9), start school — a mix of in-person classes and individualized learning supported by AI tutors.

The robots don’t teach values or replace teachers. They handle repetition, pacing, practice drills and personalized feedback. Human teachers focus on critical thinking, creativity, collaboration and emotional development.

Homework no longer means frustrated parents or late-night meltdowns. The kids work through lessons with a patient, adaptive tutor that never gets tired or irritated.


1:00 PM — Care for the Elderly, With Dignity

Mark’s mother lives nearby. She’s 82 and still independent, but needs help with mobility, medication and daily routines.

Her humanoid assistant helps her get dressed, prepares meals, reminds her to take medication, and monitors her health in real time. If anything unusual happens, Mark and Sarah are notified instantly.

She still has human carers visit for companionship and medical checkups — but the robot ensures she’s safe, supported and never alone.

This alone has transformed aging from a crisis into a manageable, dignified stage of life.


5:30 PM — Dinner Without Stress

The family arrives home.

Dinner is ready. Not frozen meals or synthetic paste, but fresh food cooked to their preferences. The robot adjusted the menu because Sarah mentioned she felt like Italian that morning.

There’s no arguing about whose turn it is to cook. No mess. No cleanup afterward.

The kitchen stays clean while the family eats together.


7:00 PM — More Time to Be Human

Instead of collapsing onto the couch exhausted, the family goes for a walk. They talk. They laugh. They play a board game.

The robot quietly handles the evening chores in the background.

This is the real revolution: time.

Time for relationships.
Time for health.
Time for creativity.
Time for rest.


The Bigger Picture

Humanoid robots didn’t replace humans. They replaced unpaid labor, repetitive work, physical strain, and logistical chaos.

They didn’t destroy jobs. They shifted them upward — toward roles that require judgment, compassion, creativity, and responsibility.

They didn’t make people lazy. They made them freer.

Cleaner homes.
Safer streets.
Better care for the elderly.
More personalized education.
Lower costs for basic services.
More time for living.


A Different Kind of Future

The future with humanoid robots isn’t Blade Runner.

It’s closer to something far more radical and far more threatening to pessimists:

A calmer, healthier, more humane society.

And when people look back at 2025 and ask why we were so afraid of machines that fold laundry, cook dinner and help grandma walk safely down the hallway, the answer will be the same as it always is.

We were afraid of change.

And we were wrong.