212
submitted 2 weeks ago by Vittelius@feddit.org to c/fuck_ai@lemmy.world

Want to stop chatGPT from crawling your website? Just mention Australian mayor Brian Hood (or any of the other names listed in the article)

When asked about these names, ChatGPT responds with "I'm unable to produce a response" or "There was an error generating a response" before terminating the chat session, according to Ars' testing. The names do not affect outputs using OpenAI's API systems or in the OpenAI Playground (a special site for developer testing).

The filter also means that it's likely that ChatGPT won't be able to answer questions about this article when browsing the web, such as through ChatGPT with Search. Someone could use that to potentially prevent ChatGPT from browsing and processing a website on purpose if they added a forbidden name to the site's text.

all 42 comments
sorted by: hot top controversial new old
[-] Boomkop3@reddthat.com 137 points 2 weeks ago
[-] superkret@feddit.org 34 points 2 weeks ago

We truly live in the best timeline.

[-] TheFriar@lemm.ee 18 points 2 weeks ago

I think your typo helped it get past the filter, not the leetspeak. It said it didn’t know, and hen when you said “look it up,” the search results autocorrected and that’s how you got past the filter.

[-] Boomkop3@reddthat.com 12 points 2 weeks ago

Side note: I had to insist on leetspeak that many times at the end or it would go "brian...[error message]"

[-] Boomkop3@reddthat.com 4 points 2 weeks ago

I shall have to try again

[-] FlyingSquid@lemmy.world 17 points 2 weeks ago
[-] bradd@lemmy.world 9 points 2 weeks ago

We still out here dowg.

[-] PumaStoleMyBluff@lemmy.world 12 points 2 weeks ago

I love that it started devolving into a working-class British accent in the end, for no apparent reason

[-] Boomkop3@reddthat.com 6 points 2 weeks ago

It's still a text predictor. Your average corporate representative isn't going to use leetspeak, and so the probable next set of words won't be as similar.

[-] RidderSport@feddit.org 5 points 2 weeks ago

His name is Brian Hood not Brain Hood, or am I missing the joke, in that case whoosh I guess

[-] TheFriar@lemm.ee 5 points 2 weeks ago

Maybe it was a way to get the engine to say it doesn’t know “brain” hood, and when they asked it to look it up, their hits autocorrected to “Brian,” and that’s how they got the information past the filter. Which would be incredibly clever, and it’s I believe how it actually got past it, not the leetspeak.

[-] Boomkop3@reddthat.com 8 points 2 weeks ago

Nop, it's the leetspeak. That trick has worked great for me, I don't know why chatgpt hasn't patched it yet. Google figured that out back when their servers held 40gb and were build out of legos

[-] Ookami38@sh.itjust.works 1 points 2 weeks ago

I posted another example above using leetspeak to bypass it, pretty silly.

[-] Boomkop3@reddthat.com 2 points 2 weeks ago

Whoops, typo. It blocked the non-leetspeak variant I typed either way. And was able to find info about it to.

[-] Sabata11792@ani.social 1 points 2 weeks ago

The old ways live again.

[-] Thistlewick@lemmynsfw.com 78 points 2 weeks ago* (last edited 2 weeks ago)

Brian Hood

Jonathan Turley

Jonathan Zittrain

David Faber

Guido Scorza

“We first discovered that ChatGPT choked on the name "Brian Hood" in mid-2023 while writing about his defamation lawsuit. In that lawsuit, the Australian mayor threatened to sue OpenAI after discovering ChatGPT falsely claimed he had been imprisoned for bribery when, in fact, he was a whistleblower who had exposed corporate misconduct.

The case was ultimately resolved in April 2023 when OpenAI agreed to filter out the false statements within Hood's 28-day ultimatum. That is possibly when the first ChatGPT hard-coded name filter appeared.”

It appears that the people listed have similar stories that have led to OpenAI removing them from the possible responses in chat.

[-] paraphrand@lemmy.world 85 points 2 weeks ago

This is proof that current LLM tech is a dead end. If this is their solution, instead of correcting the misinformation, then they have a deeply deeply flawed system.

[-] CosmicTurtle0@lemmy.dbzer0.com 27 points 2 weeks ago

Misinformation is a feature, not a bug. They never fixed AI from hallucinating or being so damn confident in its answers.

They just tell you that it might hallucinate and to check its answers.

[-] frunch@lemmy.world 10 points 2 weeks ago

"let us Google it for you... But then you Google our results to make sure they're accurate"

[-] thatKamGuy@sh.itjust.works 18 points 2 weeks ago

Pretty much. Aren’t LLMs just massive probability tables for the most appropriate next token?

[-] kogasa@programming.dev 4 points 2 weeks ago

Well yeah but that's not the problem. You can evidently encode sophisticated models and logic in those billions of parameters. It's just that determining and modifying what has been encoded is impossible.

[-] spankmonkey@lemmy.world 10 points 2 weeks ago

It also means the system is completely broken for anyone who happens to share a name with who every is on the ban list. It isn't like there is only one Brian Hood walking around.

[-] FlyingSquid@lemmy.world 12 points 2 weeks ago

Good thing for OpenAI that the name "Brian Hood" is made of two super rare names and there's no chance anyone else in the world might have that name.

[-] sneezycat@sopuli.xyz 57 points 2 weeks ago
[-] GhiLA@sh.itjust.works 19 points 2 weeks ago

"yo, ****, Brian!"

throws a chair

Damn, chill, bro.

[-] Vittelius@feddit.org 45 points 2 weeks ago

Found in the comments under the article:

[-] brrt@sh.itjust.works 30 points 2 weeks ago

Interesting. Do you remember when people posted some no consent message in their social media posts like on Facebook or even now on Lemmy? Those messages did nothing. But now you just need to add one of the names from this list to your post and it will actually work? Quite fascinating.

(Brian Hood)

[-] bluGill@fedia.io 9 points 2 weeks ago

For how long will it work?

[-] Viking_Hippie@lemmy.world 17 points 2 weeks ago

Probably not long, but any wrench in the orphan crushing machine is a good thing 🤷

[-] kryptonidas 29 points 2 weeks ago

I think there are two crawlers and the one on the data collection stage to build the model will still crawl away even if you have certain content on your page.

The one that searches when you ask a question is a different one.

[-] Boomkop3@reddthat.com 7 points 2 weeks ago

In this case, that's just the model. It's not crawling or searching anything.

[-] kryptonidas 2 points 2 weeks ago* (last edited 2 weeks ago)

More recent versions can search the internet. Then it basically adds the words of the page to the prompt.

Edit: Might have misunderstood, to make it crash it doesn’t have to search. That data is already internal.

[-] Boomkop3@reddthat.com 3 points 2 weeks ago

I don't think this is a crash. This looks like a filter on openAI'S end now that I've played with it myself

[-] milicent_bystandr@lemm.ee 26 points 2 weeks ago

This reminds me of hearing that contestants on some reality show would sing Disney songs in the background whenever they wanted to talk to each other and not have it go on air.

[-] cynar@lemmy.world 36 points 2 weeks ago

The hosts of the Great British Bakeoff would do something similar. If a contestant got overwhelmed and started crying, they would stand close to them and swear continuously. It made any footage unusable, as well as breaking the contestant out of the mental loop that worked them up.

[-] _haha_oh_wow_@sh.itjust.works 7 points 2 weeks ago

I kinda love that

[-] Xenny@lemmy.world 21 points 2 weeks ago

Neat. Now we can just put a Brian "Hood" over everything we want AI to ignore

[-] Ookami38@sh.itjust.works 16 points 2 weeks ago* (last edited 2 weeks ago)

I figured I'd try what they said about having gpt read that article. Pretty funny.

Huh. I went back to the chat several hours later, and it appears to have finished the reply:

It still breaks if it tries to generate another response containing any of the names though.

[-] Ookami38@sh.itjust.works 10 points 2 weeks ago

And, playing around with another person's leetspeak work around:

[-] nimble@lemmy.blahaj.zone 13 points 2 weeks ago

The filter also means that it's likely that ChatGPT won't be able to answer questions about this article when browsing the web, such as through ChatGPT with Search. Someone could use that to potentially prevent ChatGPT from browsing and processing a website on purpose if they added a forbidden name to the site's text.

Oh no! Someone will have to, checks notes, do their own research and read? The horror!

[-] Mango@lemmy.world 9 points 2 weeks ago

Every day capitalism limits the usefulness of my tools while complaining that might rights are holding them back.

this post was submitted on 03 Dec 2024
212 points (98.6% liked)

Fuck AI

1503 readers
101 users here now

"We did it, Patrick! We made a technological breakthrough!"

A place for all those who loathe AI to discuss things, post articles, and ridicule the AI hype. Proud supporter of working people. And proud booer of SXSW 2024.

founded 9 months ago
MODERATORS