this post was submitted on 16 Aug 2025
505 points (93.8% liked)

Technology

74180 readers
4906 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related news or articles.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] kromem@lemmy.world 0 points 1 day ago (1 children)

You do realize the majority of the training data the models were trained on was anthropomorphic data, yes?

And that there's a long line of replicated and followed up research starting with the Li Emergent World Models paper on Othello-GPT that transformers build complex internal world models of things tangential to the actual training tokens?

Because if you didn't know what I just said to you (or still don't understand it), maybe it's a bit more complicated than your simplified perspective can capture?

[–] Tattorack@lemmy.world 2 points 1 day ago (1 children)

It's not a perspective. It just is.

It's not complicated at all. The AI hype is just surrounded with heaps of wishful thinking, like the paper you mentioned (side note; do you know how many papers on string theory there are? And how many of those papers are actually substantial? Yeah, exactly).

A computer is incapable of becoming your new self aware, evolved, best friend simply because you turned Moby Dick into a bunch of numbers.

[–] kromem@lemmy.world 0 points 17 hours ago (1 children)

You do know how replication works?

When a joint Harvard/MIT study finds something, and then a DeepMind researcher follows up replicating it and finding something new, and then later on another research team replicates it and finds even more new stuff, and then later on another researcher replicates it with a different board game and finds many of the same things the other papers found generalized beyond the original scope…

That's kinda the gold standard?

The paper in question has been cited by 371 other papers.

I'm pretty comfortable with it as a citation.

[–] Tattorack@lemmy.world 1 points 16 hours ago (1 children)

Citation like that means it's a hot topic. Doesn't say anything about the quality of the research. Certainly isn't evidence of lacking bias. And considering everyone wants their AI to be the first one to be aware to some degree, everyone making claims like yours is heavily biased.

[–] kromem@lemmy.world 0 points 7 hours ago* (last edited 7 hours ago)

I'm sorry dude, but it's been a long day.

You clearly have no idea WTF you are talking about.

The research other than the DeepMind researcher's independent follow-up was all being done at academic institutions, so it wasn't "showing off their model."

The research intentionally uses a toy model to demonstrate the concept in a cleanly interpretable way, to show that transformers are capable and do build tangential world models.

The actual SotA AI models are orders of magnitude larger and fed much more data.

I just don't get why AI on Lemmy has turned into almost the exact same kind of conversations as explaining vaccine research to anti-vaxxers.

It's like people don't actually care about knowing or learning things, just about validating their preexisting feelings about the thing.

Huzzah, you managed to dodge learning anything today. Congratulations!