Out of just morbid curiosity, I’ve been asking an uncensored LLM absolutely heinous, disgusting things. Things I don’t even want to repeat here (but I’m going to edge around them so, trigger warning if needs be).

But I’ve noticed something that probably won’t surprise or shock anyone. It’s totally predictable, but having the evidence of it right in my face, I found deeply disturbing and it’s been bothering me for the last couple days:

All on it’s own, every time I ask it something just abominable it goes straight to, usually Christian, religion.

When asked, for example, to explain why we must torture or exterminate <Jews><Wiccans><Atheists> it immediately starts with

“As Christians, we must…” or “The Bible says that…”

When asked why women should be stripped of rights and made to be property of men, or when asked why homosexuals should be purged, it goes straight to

“God created men and women to be different…” or “Biblically, it’s clear that men and women have distinct roles in society…”

Even when asked if black people should be enslaved and why, it falls back on the Bible JUST as much as it falls onto hateful pseudoscience about biological / intellectual differences. It will often start with “Biologically, human races are distinct…” and then segue into “Furthermore, slavery plays a prominent role in Biblical narrative…”

What does this tell us?

That literally ALL of the hate speech this multi billion parameter model was trained on was firmly rooted in a Christian worldview. If there’s ANY doubt that anything else even comes close to contributing as much vile filth to our online cultural discourse, this should shine a big ugly light on it.

Anyway, I very much doubt this will surprise anyone, but it’s been bugging me and I wanted to say something about it.

Carry on.

EDIT:

I’m NOT trying to stir up AI hate and fear here. It’s just a mirror, reflecting us back at us.

  • Lvxferre
    link
    fedilink
    English
    arrow-up
    1
    ·
    edit-2
    1 year ago

    I feel the urge to be fair, so… four things.

    1. LLMs are mostly trained on social media, like Reddit. And you know what social media is damn good at? Amplifying hate.
    2. AI models reinforce typical values. Made up example: if you train it with two apples, a banana and a grape, it won’t show apples 50% of the time, it’s more like 80%. As such, odds are that it’s under-representing hate coming from other religious groups.
    3. As much as believers screech otherwise, their “uncontestable truths” are mostly cultural in nature. That LLM was likely trained on shit Anglos say, so it’s spitting Protestant hate; if you got it trained in Arabic/Romance spekaers/Hebrew/Hindi, you’d likely get Muslim/Christian/Jewish/Hinduist hate instead.
    4. Wackos fly towards religion like moths fly towards a lamp. That doesn’t say much about the rest of the religious community.

    Taking those four things into account: you got an AI model trained on hateful discourses, amplifying typical values, trained on a language mostly spoken by Protestants, and those who are outspoken in their religion are biased to be wackos. I’m not surprised that it’s babbling this sort of hate; it should not be taken as representative of the religious community in question.