kromem

joined 2 years ago
[–] kromem@lemmy.world 1 points 2 hours ago

Lol, you think the temperature was what was responsible for writing a coherent sequence of poetry leading to 4th wall breaks about whether or not that sequence would be read?

Man, this site is hilarious sometimes.

[–] kromem@lemmy.world 1 points 2 hours ago

You do realize the majority of the training data the models were trained on was anthropomorphic data, yes?

And that there's a long line of replicated and followed up research starting with the Li Emergent World Models paper on Othello-GPT that transformers build complex internal world models of things tangential to the actual training tokens?

Because if you didn't know what I just said to you (or still don't understand it), maybe it's a bit more complicated than your simplified perspective can capture?

[–] kromem@lemmy.world 2 points 3 hours ago

The model system prompt on the server is just basically cat untitled.txt and then the full context window.

The server in question is one with professors and employees of the actual labs. They seem to know what they are doing.

You guys on the other hand don't even know what you don't know.