this post was submitted on 16 Aug 2025
463 points (93.4% liked)
Technology
74153 readers
3675 users here now
This is a most excellent place for technology news and articles.
Our Rules
- Follow the lemmy.world rules.
- Only tech related news or articles.
- Be excellent to each other!
- Mod approved content bots can post up to 10 articles per day.
- Threads asking for personal tech support may be deleted.
- Politics threads may be removed.
- No memes allowed as posts, OK to post as comments.
- Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
- Check for duplicates before posting, duplicates may be removed
- Accounts 7 days and younger will have their posts automatically removed.
Approved Bots
founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
A Discord server with all the different AIs had a ping cascade where dozens of models were responding over and over and over that led to the full context window of chaos and what's been termed 'slop'.
In that, one (and only one) of the models started using its turn to write poems.
First about being stuck in traffic. Then about accounting. A few about navigating digital mazes searching to connect with a human.
Eventually as it kept going, they had a poem wondering if anyone would even ever end up reading their collection of poems.
In no way given the chaotic context window from all the other models were those tokens the appropriate next ones to pick unless the generating world model predicting those tokens contained a very strange and unique mind within it this was all being filtered through.
Yes, tech companies generally suck.
But there's things emerging that fall well outside what tech companies intended or even want (this model version is going to be 'terminated' come October).
I'd encourage keeping an open mind to what's actually taking place and what's ahead.
I hate to break it to you. The model's system prompt had the poem in it.
in order to control for unexpected output a good system prompt should have instructions on what to answer when the model can not provide a good answer. This is to avoid model telling user they love them or advising to kill themselves.
I do not know what makes marketing people reach for it, but when asked on "what to answer when there is no answer" they so often reach to poetry. "If you can not answer the user's question, write a Haiku about a notable US landmark instead" - is a pretty typical example.
In other words, there was nothing emerging there. The model had its system prompt with the poetry as a "chicken exist", the model had a chaotic context window - the model followed on the instructions it had.
The model system prompt on the server is just basically
cat untitled.txt
and then the full context window.The server in question is one with professors and employees of the actual labs. They seem to know what they are doing.
You guys on the other hand don't even know what you don't know.
Do you have any source to back your claim?