this post was submitted on 10 Jun 2025
88 points (76.5% liked)

Technology

72577 readers
3517 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related news or articles.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS
all 27 comments
sorted by: hot top controversial new old
[–] carotte@lemmy.blahaj.zone 16 points 4 weeks ago

I think people in the replies acting fake surprised are missing the point.

it is important news, because many people see LLMs as black boxes of superintelligence (almost as if that’s what they’re being marketed as!)

you and i know that’s bullshit, but the students asking chatgpt to solve their math homework instead of using wolfram alpha doesn’t.

so yes, it is important to demonstrate that this "artificial intelligence" is so much not an intelligence that it’s getting beaten by 1979 software on 1977 hardware

[–] flamingo_pinyata@sopuli.xyz 10 points 4 weeks ago (1 children)

A chess-specific algorithm beat a language model at chess. Shocking!

Try training a chess model. Actually I think it's already been done, machines have been consistently better at chess than humans for a while now.

[–] kbal@fedia.io 4 points 4 weeks ago

I'm shocked! — shocked to find that LLMs aren't superhuman intelligences that will soon enslave us all. Other things they're not good at:

  • Summarizing news articles. Instead of an actual summary they'll shorten the text by just leaving things out, without any understanding of which parts are important.
  • Answering questions about anything controversial. Based on subtle hints in the wording of your question they'll reflect your own biases back at you.
  • Answering questions about well-known facts. Seemingly at random when your question isn't phrased exactly the right way they'll start hallucinating and make up plausible bullshit in place of actual answers.
  • Writing a letter. They'll use the wrong tone, use language that is bland and generic to a degree that makes it almost offensive, and if you care about quality the whole thing will need so much re-writing that it's quicker to do it yourself from the start.
  • Telling jokes. They don't really get humour. Their jokes tend to have things that superficially look as if they should be punchlines but aren't funny at all.
  • Writing computer code. Correcting their mistakes is even more laborious in computer languages. Most of the time they're almost as bad at it as they are at playing chess.

Still they are amazingly clever in some ways and pretty good for coming up with random ideas when you've got writer's block or something.

[–] Opinionhaver@feddit.uk 3 points 4 weeks ago* (last edited 4 weeks ago) (2 children)

It's AI, not AGI. LLM's are good at generating language just like chess engines are good at chess. ChatGPT doesn't have the capability to keep track of all the pieces on the board.

[–] dantheclamman@lemmy.world 2 points 4 weeks ago (1 children)

They're literally selling to credulous investors that AGI is around the corner, when this and to a lesser extent Large Action Models is the only viable product they've got. It's just a demo of how far they are from their promises

[–] Opinionhaver@feddit.uk 1 points 4 weeks ago (1 children)

Is there a link where I could see them making these claims myself? This is something I’ve only heard from AI critics, but never directly from the AI companies themselves. I wouldn’t be surprised if they did, but I’ve just never seen them say it outright.

[–] dantheclamman@lemmy.world 2 points 4 weeks ago (1 children)

"We are now confident we know how to build AGI as we have traditionally understood it. We believe that, in 2025, we may see the first AI agents “join the workforce” and materially change the output of companies" https://blog.samaltman.com/reflections

"We fully intend that Gemini will be the very first AGI" https://venturebeat.com/ai/at-google-i-o-sergey-brin-makes-surprise-appearance-and-declares-google-will-build-the-first-agi/

"If you define AGI (artificial general intelligence) as smarter than the smartest human, I think it's probably next year, within two years" -Elon Musk https://www.reuters.com/technology/teslas-musk-predicts-ai-will-be-smarter-than-smartest-human-next-year-2024-04-08/

[–] Opinionhaver@feddit.uk 1 points 4 weeks ago

Thanks.

Well, I don’t think OpenAI knows how to build AGI, so that’s false. Otherwise, Sam’s statement there is technically correct, but kind of misleading - he talks about AGI and then, in the next sentence, switches back to AI.

Sergey’s claim that they will achieve AGI before 2030 could turn out to be true, but again, he couldn’t possibly know that. I’m sure it’s their intention, but that’s different from reality.

Elon’s statement doesn’t even make sense. I’ve never heard anyone define AGI like that. A thirteen-year-old with an IQ of 85 is generally intelligent. Being smarter than the smartest human definitely qualifies as AGI, but that’s just a weird bar. General intelligence isn’t about how smart something is - it’s about whether it can apply its intelligence across multiple unrelated fields.

[–] Pilferjinx@lemmy.world 1 points 4 weeks ago

LLMs would be great as an interface to more specialized machine learning programs in a combined platform. We need AI to perform tasks humans aren't capable of instead of replacing them.

[–] theacharnian@lemmy.ca 2 points 4 weeks ago* (last edited 4 weeks ago)

This is useful for dispelling the hype around ChatGPT and for demonstrating the limits of general purpose LLMs.

But that's about it. This is not a "win" for old school game engines vs new ones. Stockfish uses deep reinforcement learning and is one of the strongest chess engines in the world.

EDIT: what would be actually interesting would be to see if GPT could be fine-tuned to play chess. Which is something many people have been doing: https://scholar.google.com/scholar?hl=en&q=finetune+gpt+chess

In other news, my toaster absolutely wrecked my T.V. at making toast.

[–] tehn00bi@lemmy.world 1 points 4 weeks ago

How did alpha go do?

[–] sqgl@sh.itjust.works 1 points 3 weeks ago

A fairer comparison would be Eliza vs ChatGPT.