212
submitted 2 weeks ago by Vittelius@feddit.org to c/fuck_ai@lemmy.world

Want to stop chatGPT from crawling your website? Just mention Australian mayor Brian Hood (or any of the other names listed in the article)

When asked about these names, ChatGPT responds with "I'm unable to produce a response" or "There was an error generating a response" before terminating the chat session, according to Ars' testing. The names do not affect outputs using OpenAI's API systems or in the OpenAI Playground (a special site for developer testing).

The filter also means that it's likely that ChatGPT won't be able to answer questions about this article when browsing the web, such as through ChatGPT with Search. Someone could use that to potentially prevent ChatGPT from browsing and processing a website on purpose if they added a forbidden name to the site's text.

you are viewing a single comment's thread
view the rest of the comments
[-] Thistlewick@lemmynsfw.com 78 points 2 weeks ago* (last edited 2 weeks ago)

Brian Hood

Jonathan Turley

Jonathan Zittrain

David Faber

Guido Scorza

“We first discovered that ChatGPT choked on the name "Brian Hood" in mid-2023 while writing about his defamation lawsuit. In that lawsuit, the Australian mayor threatened to sue OpenAI after discovering ChatGPT falsely claimed he had been imprisoned for bribery when, in fact, he was a whistleblower who had exposed corporate misconduct.

The case was ultimately resolved in April 2023 when OpenAI agreed to filter out the false statements within Hood's 28-day ultimatum. That is possibly when the first ChatGPT hard-coded name filter appeared.”

It appears that the people listed have similar stories that have led to OpenAI removing them from the possible responses in chat.

[-] paraphrand@lemmy.world 85 points 2 weeks ago

This is proof that current LLM tech is a dead end. If this is their solution, instead of correcting the misinformation, then they have a deeply deeply flawed system.

[-] CosmicTurtle0@lemmy.dbzer0.com 27 points 2 weeks ago

Misinformation is a feature, not a bug. They never fixed AI from hallucinating or being so damn confident in its answers.

They just tell you that it might hallucinate and to check its answers.

[-] frunch@lemmy.world 10 points 2 weeks ago

"let us Google it for you... But then you Google our results to make sure they're accurate"

[-] thatKamGuy@sh.itjust.works 18 points 2 weeks ago

Pretty much. Aren’t LLMs just massive probability tables for the most appropriate next token?

[-] kogasa@programming.dev 4 points 2 weeks ago

Well yeah but that's not the problem. You can evidently encode sophisticated models and logic in those billions of parameters. It's just that determining and modifying what has been encoded is impossible.

[-] spankmonkey@lemmy.world 10 points 2 weeks ago

It also means the system is completely broken for anyone who happens to share a name with who every is on the ban list. It isn't like there is only one Brian Hood walking around.

[-] FlyingSquid@lemmy.world 12 points 2 weeks ago

Good thing for OpenAI that the name "Brian Hood" is made of two super rare names and there's no chance anyone else in the world might have that name.

this post was submitted on 03 Dec 2024
212 points (98.6% liked)

Fuck AI

1503 readers
103 users here now

"We did it, Patrick! We made a technological breakthrough!"

A place for all those who loathe AI to discuss things, post articles, and ridicule the AI hype. Proud supporter of working people. And proud booer of SXSW 2024.

founded 9 months ago
MODERATORS