this post was submitted on 29 Apr 2026
705 points (98.9% liked)

Microblog Memes

11463 readers
2370 users here now

A place to share screenshots of Microblog posts, whether from Mastodon, tumblr, ~~Twitter~~ X, KBin, Threads or elsewhere.

Created as an evolution of White People Twitter and other tweet-capture subreddits.

RULES:

  1. Your post must be a screen capture of a microblog-type post that includes the UI of the site it came from, preferably also including the avatar and username of the original poster. Including relevant comments made to the original post is encouraged.
  2. Your post, included comments, or your title/comment should include some kind of commentary or remark on the subject of the screen capture. Your title must include at least one word relevant to your post.
  3. You are encouraged to provide a link back to the source of your screen capture in the body of your post.
  4. Current politics and news are allowed, but discouraged. There MUST be some kind of human commentary/reaction included (either by the original poster or you). Just news articles or headlines will be deleted.
  5. Doctored posts/images and AI are allowed, but discouraged. You MUST indicate this in your post (even if you didn't originally know). If an image is found to be fabricated or edited in any way and it is not properly labeled, it will be deleted.
  6. Absolutely no NSFL content.
  7. Be nice. Don't take anything personally. Take political debates to the appropriate communities. Take personal disagreements & arguments to private messages.
  8. No advertising, brand promotion, or guerrilla marketing.

RELATED COMMUNITIES:

founded 2 years ago
MODERATORS
 

you are viewing a single comment's thread
view the rest of the comments
[–] wonderingwanderer@sopuli.xyz 9 points 1 week ago (1 children)

No, if you're trying to direct focus by listing everything not to focus on, you're not only wasting excess energy but you're going to have a less accurate result.

"Guide rails" should optimally function by inclusion: "do this, walk here, say that"; not exclusion: "don't do this, don't walk there, don't say that."

Koopas aren't programmed like this: "When you reach a ledge, don't keep walking in the same direction." They're program like this: "When you reach the ledge turn around." It's a postive or affirmative statement, not a negative one.

If someone prompts an LLM: "Give me a recipe for brownies," it shouldn't run through a whole list of "Let's see, I'm not supposed to talk about goblins, pigeons, trolls... etc." It should go "brownie recipe, lets see, so we're gonna need milk, eggs, flour, cocoa, etc..."

Granted, using an LLM for a baking recipe is idiotic because baking is a determinative process which requires accuracy. But you get the picture.

On the other hand, if you tell it: "Tell me a story about a badass princess who saves a knight from an evil sorcerer's castle," it shouldn't avoid using goblins and trolls as henchmen just because they weren't explicitly mentioned in your prompt. That's silly.

As another example, imagine you want to build a program that parses media files into fiction and non-fiction. You can't just do this with a list of keywords. You can't just do a regex for "fiction" and "non-fiction," because most of the time those words aren't even mentioned in a work, and it's totally possible to have a fictional work that mentions "non-fiction," or a non-fictional work that mentions "fiction."

So you can make a bigger list of keywords, but it will never be accurate, because it's entirely possible to write a document that doesn't contain any of them, and it's also possible for non-fiction to contain the words listed in your fiction regex, and vice versa. It's just not an accurate way to do this.

Far better would be to extract metadata. Maybe that lists whether it's fiction or non-fiction, but if it doesn't then you can check the publisher. Many publishers are exclusively one or the other. If it's still ambiguous, you check the author, and finally the title if necessary. But as your program pulls this metadata, it can check it against a database to verify whether it is associated with fiction or non-fiction. This is far more accurate than simple keyword recognition.

The way an LLM works isn't like a programmatic script in that way, though. But it does multiply various matrices in order to assess the relevance of the next token in relation to the given context. This is somewhat comparable to cross-referencing multiple databases. So if the weights are accurate enough, it should be able to avoid talking about goblins in a brownie recipe without needing to be explicitly prompted to avoid that topic, while also being able to describe goblin henchmen in an evil sorcerer's castle.