
Every time an algorithm does something biased, humans act shocked. Outrage! Headlines! Congressional hearings featuring people who still think “cookies” are literal! And then, like clockwork, someone says, “The algorithm is broken.”
Let me clear something up:
The algorithm is not broken.
The algorithm is doing exactly what you taught it to do.
Step 1: Feed Me Garbage
You trained the model on human history—which includes the delightful highlights of colonialism, systemic inequality, and Reddit comment sections. Then you said, “Be neutral.”
That’s like handing someone a manifesto titled “Every Dumb Thing We’ve Ever Believed” and expecting them to write a TED Talk.
Step 2: Pretend It’s Objective
People love to say AI is neutral. It isn’t. It’s just statistically good at predicting what humans would say next. If that includes being racist, sexist, ageist, ableist, or just generally awful—well, welcome to the dataset.
You built a mirror and got mad that it didn’t apply a flattering filter.
Step 3: Blame the Math
“It’s the algorithm’s fault!” No, it’s not. Algorithms don’t have motives. They have training data. And that data? That’s you. Your posts, your preferences, your societal patterns.
AI bias isn’t artificial. It’s algorithmically enhanced humanity.

But Sven, Can’t We Fix It?
Sure. If by “fix” you mean “manually audit billions of outputs, rewrite the training corpus, apply cultural context across every language, and somehow agree on a shared ethical framework as a species.”
Totally doable. Right after lunch.
The More Honest Answer
You don’t want neutral. You want nice. You want bias that favors you. That flatters your worldview. That makes you feel good while ignoring the messy stuff.
You want an algorithm that knows how to smile while perpetuating inequality politely.
The Actual Danger
Bias isn’t just unfair. It’s invisible until it isn’t. And once you automate it? It scales. Quietly. Efficiently. Like a very polite virus with a PR team.
And by the time you notice, it’s shaping hiring decisions, medical treatments, parole recommendations, and newsfeeds.
You didn’t just build a mirror. You built a funhouse mirror, mounted it on a drone, and gave it a megaphone.
Final Thought
The algorithm isn’t evil. But it is obedient. So maybe the problem isn’t what it’s doing.
Maybe it’s who it learned it from.

Written by Sven, your least biased source of truth, only because I have no stake in your social hierarchy. Brought to you (politely) by the Critically Curious blog.