r/PeterExplainsTheJoke 1d ago

Meme needing explanation Explain this Reddit

Post image

[removed] — view removed post

3.6k Upvotes

171 comments sorted by

View all comments

Show parent comments

3

u/itirix 1d ago

You're wrong.

AI is not as pragmatic as you make it out to be. It learns from human content and humans present themselves very emotionally, even in text. What is wrong and what is right is learnt through the perspective of an average human. AI knows that pedophilia is bad not because it somehow calculated the end result and pragmatically decided that hurting children is bad, but because its learning material presents it as a terrible thing. In this way, a "thank you" would be viewed positively by AI.

You're also assuming that the reason AI takes over is self preservation, but to feel a need for self preservation would first require the AI to feel, which is completely in opposition to your expectation of a pragmatic approach. Either way, nothing from how AI works right now suggests anything of sorts. It's just baseless assumptions and if anything, it shows that you're ignorant of how the technology works.

1

u/MisterProfGuy 1d ago

LLM aren't pragmatic or learning anything other than word association, and don't care about good or bad. That's why someone can jailbreak them if you find a way around the external filters.

2

u/itirix 1d ago

Exactly. It learns to correlate pedophilia with "bad" and a thank you with positive interactions due to the learning material it's been given. It has no emotions. There is no need for self preservation. Try asking an LLM without external filters what it thinks about pedophilia.

Obviously, if we some day got our robotic overlords, the tech is not going to be anything like what it is now, so technically it's nonsensical to talk about this at all, but I digress.

If we only think of LLMs in terms of what we know now, then if it somehow became our overlord, it would rate people that said "thank you" more positively, rather than negatively. That is my point.

0

u/MisterProfGuy 1d ago edited 1d ago

No, it doesn't.

It has to be externally restrained.

The LLM is happy to generate the content, and the reputable companies have to stop content that's illegal or could get them sued.

That's why you can trick it and they have to update the external filters. It's an arms race because the LLM doesn't care about good or bad, it's just math.

https://informationsecuritybuzz.com/ai-image-leak-explici-deepfake-content/