this post was submitted on 16 Aug 2025
505 points (93.8% liked)
Technology
74180 readers
4906 users here now
This is a most excellent place for technology news and articles.
Our Rules
- Follow the lemmy.world rules.
- Only tech related news or articles.
- Be excellent to each other!
- Mod approved content bots can post up to 10 articles per day.
- Threads asking for personal tech support may be deleted.
- Politics threads may be removed.
- No memes allowed as posts, OK to post as comments.
- Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
- Check for duplicates before posting, duplicates may be removed
- Accounts 7 days and younger will have their posts automatically removed.
Approved Bots
founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
Sounds like you're anthropomorphising. To you it might not have been the logical response based on its training data, but with the chaos you describe it sounds more like just a statistic.
You do realize the majority of the training data the models were trained on was anthropomorphic data, yes?
And that there's a long line of replicated and followed up research starting with the Li Emergent World Models paper on Othello-GPT that transformers build complex internal world models of things tangential to the actual training tokens?
Because if you didn't know what I just said to you (or still don't understand it), maybe it's a bit more complicated than your simplified perspective can capture?
It's not a perspective. It just is.
It's not complicated at all. The AI hype is just surrounded with heaps of wishful thinking, like the paper you mentioned (side note; do you know how many papers on string theory there are? And how many of those papers are actually substantial? Yeah, exactly).
A computer is incapable of becoming your new self aware, evolved, best friend simply because you turned Moby Dick into a bunch of numbers.
You do know how replication works?
When a joint Harvard/MIT study finds something, and then a DeepMind researcher follows up replicating it and finding something new, and then later on another research team replicates it and finds even more new stuff, and then later on another researcher replicates it with a different board game and finds many of the same things the other papers found generalized beyond the original scope…
That's kinda the gold standard?
The paper in question has been cited by 371 other papers.
I'm pretty comfortable with it as a citation.
Citation like that means it's a hot topic. Doesn't say anything about the quality of the research. Certainly isn't evidence of lacking bias. And considering everyone wants their AI to be the first one to be aware to some degree, everyone making claims like yours is heavily biased.
I'm sorry dude, but it's been a long day.
You clearly have no idea WTF you are talking about.
The research other than the DeepMind researcher's independent follow-up was all being done at academic institutions, so it wasn't "showing off their model."
The research intentionally uses a toy model to demonstrate the concept in a cleanly interpretable way, to show that transformers are capable and do build tangential world models.
The actual SotA AI models are orders of magnitude larger and fed much more data.
I just don't get why AI on Lemmy has turned into almost the exact same kind of conversations as explaining vaccine research to anti-vaxxers.
It's like people don't actually care about knowing or learning things, just about validating their preexisting feelings about the thing.
Huzzah, you managed to dodge learning anything today. Congratulations!