this post was submitted on 16 Aug 2025
43 points (83.1% liked)

Technology

74073 readers
2818 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related news or articles.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS
all 11 comments
sorted by: hot top controversial new old
[–] A_norny_mousse@feddit.org 5 points 4 hours ago* (last edited 4 hours ago)

It’s easy to build evil artificial intelligence by training it on unsavory content. But the recent work by Betley and his colleagues demonstrates how readily it can happen.

Garbage in, garbage out.

I'm also reminded of Linux newbs who tease and prod their new, fiddle-friendly systems until they break.

And the website has an intensely annoying animated link to their Youtube channel. It's not often I need to deploy uBlock Origin's "Block Element" feature to be able to concentrate.

[–] frongt@lemmy.zip 43 points 7 hours ago (2 children)

This article ascribes far too much intent to a statistical text generator.

[–] Supervisor194@lemmy.world 4 points 2 hours ago

It is Schroedinger's Stochastic Parrot. Simultaneously a Chinese Room and the reincarnation of Hitler.

It exposes that there might be a link between bad developers and far right extremism though.

... which we already knew from Notch.

[–] Preventer79@sh.itjust.works 11 points 8 hours ago* (last edited 8 hours ago) (2 children)

Anyone know how to get access to these "evil" models?

[–] Cherry@piefed.social 1 points 1 hour ago

Access to view the evil models or to make more evil models?

[–] renegadespork@lemmy.jelliefrontier.net 8 points 7 hours ago (1 children)
[–] neinhorn@lemmy.ca 3 points 5 hours ago

Just ask Anakin

[–] kassiopaea@lemmy.blahaj.zone 8 points 8 hours ago (1 children)

I'd like to see similar testing done comparing models where the "misaligned" data is present during training, as opposed to fine-tuning. That would be a much harder thing to pull off, though.

[–] sleep_deprived@lemmy.dbzer0.com 3 points 5 hours ago

It isn't exactly what you're looking for, but you may find this interesting, and it's a bit of an insight into the relationship between pretraining and fine tuning: https://arxiv.org/pdf/2503.10965