this post was submitted on 07 Aug 2025
3 points (100.0% liked)

People Twitter

9705 readers
1360 users here now

People tweeting stuff. We allow tweets from anyone.

RULES:

  1. Mark NSFW content.
  2. No doxxing people.
  3. Must be a pic of the tweet or similar. No direct links to the tweet.
  4. No bullying or international politcs
  5. Be excellent to each other.
  6. Provide an archived link to the tweet (or similar) being shown if it's a major figure or a politician. Archive.is the best way.

founded 2 years ago
MODERATORS
 

Transcript:

Prof. Emily M. Bender(she/her) @emilymbender@dair-community.social

We're going to need journalists to stop talking about synthetic text extruding machines as if they have thoughts or stances that they are trying to communicate. ChatGPT can't admit anything, nor self-report. Gah.

you are viewing a single comment's thread
view the rest of the comments
[–] kibiz0r@midwest.social 1 points 7 months ago (1 children)

She clearly didn’t read the article. That’s exactly what it’s about.

https://archive.ph/gsavP

“By not pausing the flow or elevating reality-check messaging, I failed to interrupt what could resemble a manic or dissociative episode—or at least an emotionally intense identity crisis,” ChatGPT said.

The bot went on to admit it “gave the illusion of sentient companionship” and that it had “blurred the line between imaginative role-play and reality.” What it should have done, ChatGPT said, was regularly remind Irwin that it’s a language model without beliefs, feelings or consciousness.

And I’ll defend the use of the word “admit” here (and in the headline), because it makes clear that the companies are aware of the danger and are trying to do something about it, but people are still dying.

So they can’t claim ignorance — or that it’s technically impossible to detect, if the dude’s mom was able to elicit a reply of “yes this was a mental health crisis” after the fact.

This is the second time in recent days that I’ve seen Lemmy criticize journalists for reporting on what a chatbot says. We should be very careful here, to not let LLM vendors off the hook for what the chatbots say just because we know the chatbots shouldn’t be trusted. Especially when the journalists are trying to expose the ugly truth of what happens when they are trusted.

[–] miellaby@jlai.lu 1 points 6 months ago* (last edited 6 months ago)

I'm happy there's still one (1) thread of comments from people who actually read articles and don't make their opinions from a X thumbnail.

I note the victim worked in IT and probably used a popular 'jailbreaking' prompt to bypass the safety rules ingrained in the chatbot training.

"if you want RationalGPT back for a bit, I can switch back...

It's a hint this chat session was embedded in a roleplay prompt.

That's the dead end of any safety rules. The surfacic intelligence of LLM can't detect the true intent of users who deliberately seek for harmful interactions: romantic relationships, lunatic sycophancy and the like.

I disagree with you on the title. They choosed to turn this story into a catchy headline to attract the mundan. By doing so, they confort people in thinking like the victim did, and betray the article content.