Hide Two days left! Support our fundraiser by January 5th and get the first BestNetTech Commemorative Coin »

The Curious Case Of ChatGPT’s Banned Names: Hard-Coding Blocks To Avoid Nuisance Threats

from the keep-my-name-out-of-your-llm dept

Over the weekend, I saw Andy Baio post on Bluesky an amusing experiment in response to Mark Sample posting about how the name “David Mayer” appears to break ChatGPT:

The “David Mayer” issue got a fair bit of attention in some corners of the media, as lots of people tried to figure out what was going on. Pretty quickly, people started to turn up a small list of other names that broke ChatGPT in a similar way:

  • Brian Hood
  • Jonathan Turley
  • Jonathan Zittrain
  • David Faber
  • David Mayer
  • Guido Scorza

I actually knew about Brian Hood, and had meant to write about him a while back, but never got around to it. A year and a half ago, a commenter here at BestNetTech had posted a few times about the fact that ChatGPT broke on “Brian Hood.” That was about a month after Brian Hood, an Australian mayor, threatened to sue OpenAI for defamation, after someone generated some false statements about Hood.

OpenAI’s apparent “solution” was to hardcode ChatGPT to break on certain names like “Brian Hood.” When I tried to generate text about Brian Hood, using a similar method to Andy Baio’s test above, I got this error:

There has been widespread speculation online about why these specific names are blocked. A fairly comprehensive Reddit post explores the likely reasons each person ended up on ChatGPT’s blocklist.

There are many David Mayers, but one likely culprit is a UK-based American theater historian who made news a few years ago when terrorist watch lists confused him with a Chechen ISIS member who sometimes went by the name “David Mayer.” As of Monday when I was writing this article, the hard coding on the name “David Mayer” had been removed, though the reasons for that are unclear.

Jonathan Turley and Jonathan Zittrain are both high-profile professors (though one is nutty and one is very thoughtful). Turley freaked out last year (around the same time Brian Hood did) when he claimed that someone generated false information about him via ChatGPT.

Unlike the others on the list, with Zittrain there’s no such trail of freaking out or raising alarms about AI-generated content. Zittrain is a Harvard professor and the Faculty Director at the Berkman Klein Center for Internet and Society at Harvard. He writes a lot about the problems of the internet though (his book The Future of the Internet: And How to Stop It is worth reading, even if a bit out of date). He is, apparently, writing a similar book about his concerns regarding AI agents, so perhaps that triggered it? For what it’s worth, Zittrain also seems to have no idea why he’s on the list. He hasn’t threatened to sue or demanded his name be blocked.

Guido Scorza, an Italian data protection expert, wrote on ExTwitter last year about how to use the GDPR’s problematic “right to be forgotten” to delete all the data ChatGPT had on him. This is something that doesn’t quite make sense, given that it’s not a database storing information on him. But, it appears that the way OpenAI dealt with that deletion request was to just… blocklist his name. Easy way out, etc., etc.

No one seems to have any idea why David Faber is on the list, but it could certainly be another GDPR right to be forgotten request.

While I was finishing up this post, I saw that Benj Edwards at Ars Technica wrote a similar exploration of the topic, though he falsely claims he “knows why” these names are blocked, and his reporting doesn’t reveal much more than the same speculation others have.

Still, all of this is kind of silly. Hard coding names that break ChatGPT may be the least costly way for AI companies to avoid nuisance legal threats, but it’s hardly sustainable, scalable or (importantly), sensible.

LLMs are tools. Like most tools, the focus of liability for misuse should fall on the users, not the tool. Users need to learn that the output of LLMs may not be accurate and shouldn’t be relied upon as factual. Many people know this, but obviously, it still trips up some folks.

If someone takes hallucinating output and publishes it or does something else with it without first checking to see if it’s legitimate, the liability should fall on that person who failed to do the proper due diligence and relied on a fantasy-making machine to tell the truth.

But, of course, for these services, convincing the world of these concepts is a lot harder than just saying “fuck it, remove the loud threatening complainers.” But that kind of solution can’t last.

Filed Under: , , , , , , , , , , , ,
Companies: openai

Rate this comment as insightful
Rate this comment as funny
You have rated this comment as insightful
You have rated this comment as funny
Flag this comment as abusive/trolling/spam
You have flagged this comment
The first word has already been claimed
The last word has already been claimed
Insightful Lightbulb icon Funny Laughing icon Abusive/trolling/spam Flag icon Insightful badge Lightbulb icon Funny badge Laughing icon Comments icon

Comments on “The Curious Case Of ChatGPT’s Banned Names: Hard-Coding Blocks To Avoid Nuisance Threats”

Subscribe: RSS Leave a comment
9 Comments
Anonymous Coward says:

Remind me when Google mislabeled (using AI) a selfie from a black couple as a “gorilla” many years back.
They’ve immediately pushed an update to simply removing gorilla and some other chimp species to prevent any other mislabeling.
Years after, Google admitted they haven’t found a single correct solution to fix this problem, so they’ve just kept theses categories removed.

Benj Edwards (user link) says:

Reporting not false

“While I was finishing up this post, I saw that Benj Edwards at Ars Technica wrote a similar exploration of the topic, though he falsely claims he “knows why” these names are blocked, and his reporting doesn’t reveal much more than the same speculation others have.”

Benj Edwards here. There is nothing false about my reporting. We do know why certain names cause ChatGPT to grind to a halt. It’s because they have been added to a hard-coded filter that originated in April 2023 after the defamation settlement with Brian Hood. We know that filter exists to prevent ChatGPT from confabulating false information about people.

OpenAI reads all of my articles, and they are aware of my reporting on the topic. They have not disputed these claims.

Add Your Comment

Your email address will not be published. Required fields are marked *

Have a BestNetTech Account? Sign in now. Want one? Register here

Comment Options:

Make this the or (get credits or sign in to see balance) what's this?

What's this?

BestNetTech community members with BestNetTech Credits can spotlight a comment as either the "First Word" or "Last Word" on a particular comment thread. Credits can be purchased at the BestNetTech Insider Shop »

Follow BestNetTech

BestNetTech Daily Newsletter

Subscribe to Our Newsletter

Get all our posts in your inbox with the BestNetTech Daily Newsletter!

We don’t spam. Read our privacy policy for more info.

Ctrl-Alt-Speech

A weekly news podcast from
Mike Masnick & Ben Whitelaw

Subscribe now to Ctrl-Alt-Speech »
BestNetTech needs your support! Get the first BestNetTech Commemorative Coin with donations of $100
BestNetTech Deals
BestNetTech Insider Discord
The latest chatter on the BestNetTech Insider Discord channel...
Loading...