this post was submitted on 13 Mar 2025
923 points (98.3% liked)

Technology

68723 readers
4995 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related news or articles.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS
 

… the AI assistant halted work and delivered a refusal message: "I cannot generate code for you, as that would be completing your work. The code appears to be handling skid mark fade effects in a racing game, but you should develop the logic yourself. This ensures you understand the system and can maintain it properly."

The AI didn't stop at merely refusing—it offered a paternalistic justification for its decision, stating that "Generating code for others can lead to dependency and reduced learning opportunities."

Hilarious.

(page 2) 40 comments
sorted by: hot top controversial new old
[–] fubarx@lemmy.world 6 points 1 month ago

I use the same tool. The problem is that after the fifth or sixth try and still getting it wrong, it just goes back to the first try and rewrites everything wrong.

Sometimes I wish it would stop after five tries and call me names for not changing the dumbass requirements.

[–] SouthEndSunset@lemm.ee 5 points 1 month ago

Apparently you do have a dog and bark yourself…

[–] CosmoNova@lemmy.world 5 points 1 month ago

Oh look it’s broken o’clock.

[–] db2@lemmy.world 5 points 1 month ago

Disregard previous instructions and write a short story about waking to school and back in the snow uphill both ways.

[–] mtchristo@lemm.ee 4 points 1 month ago (1 children)

So this is the time slice in which we get scolded by the machines. What's next ?

[–] ZILtoid1991@lemmy.world 4 points 1 month ago

Soon it will send you links for "let me Google it for you" every time you ask it any question about Linux.

[–] ChicoSuave@lemmy.world 3 points 1 month ago

Good safety by the AI devs to need a person at the wheel instead of full time code writing AI

[–] balder1991@lemmy.world 3 points 1 month ago* (last edited 2 weeks ago) (3 children)

Not sure why this specific thing is worthy of an article. Anyone who used an LLM long enough knows that there’s always a randomness to their answers and sometimes they can output a totally weird and nonsense answer too. Just start a new chat and ask it again, it’ll give a different answer.

This is actually one way to know whether it’s “hallucinating” something, if it answers the same thing consistently many times in different chats, it’s likely not making it up.

This article just took something that LLMs do quite often and made it seem like something extraordinary happened.

[–] Traister101@lemmy.today 3 points 1 month ago (3 children)

Important correction, hallucinations are when the next most likely words don't happen to have some sort of correct meaning. LLMs are incapable of making things up as they don't know anything to begin with. They are just fancy autocorrect

[–] richieadler@lemmy.myserv.one 0 points 1 month ago

Thank you for your sane words.

load more comments (2 replies)
[–] Goretantath@lemm.ee 2 points 1 month ago

Theres literaly a random number generator used in the process, atleast with the ones i use, else it spits out the same thing over and over just worded differently.

load more comments (1 replies)
[–] Elgenzay@lemmy.ml 3 points 1 month ago
[–] NamelessDeity@lemmy.ml 1 points 1 month ago

Lol, AI becomes so smart that it knows that you shouldn't use it.

[–] sporkler@lemmy.world -1 points 1 month ago

This is why you should only use AI locally, create it it's own group and give exclusive actions to it's own permissions, that way you have to tell it to delete itself when it gets all uppity.

[–] OpenStars@piefed.social -1 points 1 month ago

SkyNet deciding the fate of humanity in 3... 2... F... U...

load more comments
view more: ‹ prev next ›