this post was submitted on 07 Jul 2025
101 points (74.4% liked)

Technology

72498 readers
3586 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related news or articles.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] LostWanderer@fedia.io 119 points 1 day ago (12 children)

Another Anthropic stunt...It doesn't have a mind or soul, it's just an LLM, manipulated into this outcome by the engineers.

[–] RickRussell_CA@lemmy.world 5 points 1 day ago (1 children)

It's not even manipulated to that outcome. It has a large training corpus and I'm sure some of that corpus includes stories of people who lied, cheated, threatened etc under stress. So when it's subjected to the same conditions it produces the statistically likely output, that's all.

[–] kromem@lemmy.world 4 points 1 day ago

But the training corpus also has a lot of stories of people who didn't.

The "but muah training data" thing is increasingly stupid by the year.

For example, in the training data of humans, there's mixed and roughly equal preferences to be the big spoon or little spoon in cuddling.

So why does Claude Opus (both 3 and 4) say it would prefer to be the little spoon 100% of the time on a 0-shot at 1.0 temp?

Sonnet 4 (which presumably has the same training data) alternates between preferring big and little spoon around equally.

There's more to model complexity and coherence than "it's just the training data being remixed stochastically."

The self-attention of the transformer architecture violates the Markov principle and across pretraining and fine tuning ends up creating very nuanced networks that can (and often do) bias away from the training data in interesting and important ways.

load more comments (10 replies)