this post was submitted on 28 Feb 2026
388 points (96.2% liked)
Technology
82015 readers
3919 users here now
This is a most excellent place for technology news and articles.
Our Rules
- Follow the lemmy.world rules.
- Only tech related news or articles.
- Be excellent to each other!
- Mod approved content bots can post up to 10 articles per day.
- Threads asking for personal tech support may be deleted.
- Politics threads may be removed.
- No memes allowed as posts, OK to post as comments.
- Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
- Check for duplicates before posting, duplicates may be removed
- Accounts 7 days and younger will have their posts automatically removed.
Approved Bots
founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
hi friends i hope you're well.
i worked a laborious job and experienced a phenomenon i refer to as "parasitic thought:" it is where someone will provide to you all of the information that a person would require to reach the correct conclusion, and then stare at you. they want you to crunch the info for them.
i feel like one of those parasites in my agent interactions. i know i COULD think, but you can do it too, lil buddy. go on. do it for me.
i don't know about "reasonable" or "ethical" or "polite," but in my experience: if someone just regurgitates some clank clank slop slop, it reads as hostile. "i can't be bothered to communicate with you, here, read this wall of gpt-vomit"
my instinct is to copy and paste, "LLM agent of my choice, what's this person trying to say to me?" and then skim the ai synthesized summary of the ai composed body text generated from some idiot's faint echoes of thought.
in the words of your highschool biology teacher, the human is the powerhouse of the agentic loop. in my unimportant opinion, responsible use of genai agents means that the output should be indistinguishable, if not better, than something you wrote by hand.
there are privacy implications. linguistic assessment can be used to identify you. from a privacy perspective, the internet would be preferable if everyone fed their carefully formed thoughts to an LLM and said "make this look like chatgpt 3 wrote it."