342
this post was submitted on 27 Aug 2025
342 points (97.0% liked)
Technology
74519 readers
4820 users here now
This is a most excellent place for technology news and articles.
Our Rules
- Follow the lemmy.world rules.
- Only tech related news or articles.
- Be excellent to each other!
- Mod approved content bots can post up to 10 articles per day.
- Threads asking for personal tech support may be deleted.
- Politics threads may be removed.
- No memes allowed as posts, OK to post as comments.
- Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
- Check for duplicates before posting, duplicates may be removed
- Accounts 7 days and younger will have their posts automatically removed.
Approved Bots
founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
I agree. However I do realize, like in this specific case, requiring a mandated reporter for a jailbroken prompt, given the complexity of human language, would be impossible.
Arguably, you'd have to train an entirely separate LLM to detect anything remotely considered harmful language, and the way they train their model it is not possible.
The technology simply isn't ready to use, and people are vastly unaware of how this AI works.
I fully agree. LLMs create situations that our laws aren't prepared for and we can't reasonably get them into a compliant state on account of how the technology works. We can't guarantee that an LLM won't lose coherence to the point of ignoring its rules as the context grows longer. The technology inherently can't make that kind of guarantee.
We can try to add patches like a rules-based system that scans chats and flags them for manual review if certain terms show up but whether those patches suffice will have to be seen.
Of course most of the tech industry will instead clamor for an exception because "AI" (read: LLMs and image generation) is far too important to let petty rules hold back progress. Why, if we try to enforce those rules, China will inevitably develop Star Trek-level technology within five years and life as we know it will be doomed. Doomed I say! Or something.
Hey, Star-Trek technology did enable an actual utopia, so the modern corpofascist way of life would die kicking and screaming.