this post was submitted on 20 Mar 2026
114 points (99.1% liked)

Fuck AI

6318 readers
742 users here now

"We did it, Patrick! We made a technological breakthrough!"

A place for all those who loathe AI to discuss things, post articles, and ridicule the AI hype. Proud supporter of working people. And proud booer of SXSW 2024.

AI, in this case, refers to LLMs, GPT technology, and anything listed as "AI" meant to increase market valuations.

founded 2 years ago
MODERATORS
 

cross-posted from: https://piefed.social/c/opensource/p/1899754/i-prompt-injected-my-contributing-md-50-of-prs-are-bots

How a hidden prompt injection in CONTRIBUTING.md revealed that 40% of pull requests to a popular GitHub repository were generated by AI bots

top 15 comments
sorted by: hot top controversial new old
[–] Wispy2891@lemmy.world 1 points 12 hours ago

So, this guy maintains a repo that is 100% exclusively targeting ai bros (MCP servers), then does a shocked Pikachu face when the ai bros use ai bots to contribute to the ai slop??

[–] cloudskater@piefed.blahaj.zone 24 points 3 days ago (1 children)

So this person prompt injected to figure out which submissions were bots, found out, and instead of taking a firm stance against this practice they're gonna... try and "make use" of all the bot submissions... sighhhhh oh my fucking God how many more times am I going to see these similar brain dead arguments? I can't take it anymore.

[–] WillowBe@lemmy.blahaj.zone 4 points 2 days ago

that's not what they said

[–] november@piefed.blahaj.zone 29 points 3 days ago (2 children)

What is MCP?
MCP is an open protocol that enables AI models to securely interact with local and remote resources through standardized server implementations.

Is this a different definition of AI than the slopbots we all know and love, or is this a case of poetic irony?

[–] MalReynolds@slrpnk.net 32 points 3 days ago (1 children)

MCP is for tool calling, the stochastic parrot matches something, say 133 x 8 / 7 as a math problem and instead of guessing wrong calls a calculator using good old deterministic code and gets a correct answer.

[–] Thorry@feddit.org 28 points 3 days ago (2 children)

Which has lead to a whole new issue popping up where the AI chooses not to use the tool, because it doesn't trust it. Now this is anthropomorphizing a lot, in reality it's a reward misalignment issue, but still. It's called tool aversion and can be an issue. The LLM would even make up a reason the tool won't work, or tell you it did use the tool when it didn't actually.

In order to fix this a technique called cryptographic receipts is used. This adds an expected hash output based on the tool and the input. The tool outputs the correct hash, but if the LLM didn't call the tool and made up some BS, the hash isn't there. On the router level the missing or incorrect hash is easily detected and an exception raised in order to not present the result to the user (and hopefully correct the issue).

Another whole issue with using tools for certain stuff is it requires the situation to be properly evaluated for the tool to be called in the first place. So a user might hear the marketing say: Our old AI used to mess up maths, but our latest and greatest model is super duper smart and can do maths. The user then goes to verify this is the case with some simple tests, and sees it's correct. Then follows up with some harder math problems and the output is still correct. However the user doesn't know and isn't informed the AI used the tool, the user is lead to believe the AI is smart and can do maths now. So in situations where the detection doesn't work, or the LLM doesn't use the tool for whatever reason, the user is fed a nonsense response. Which the user will fully trust, given what he knows and tested himself.

It's so annoying these AI companies continually do shit like this. Lie to users, keep them in the dark and overpromise. Users relate the AI to human intelligence. And for humans, if you can do hard math problems, the circumstances don't really matter. We grasp the basic concepts and can execute to get a result and know how to double check. For these AI tools, it might get a PhD level math question totally right, faster and better than a human. And then get a simple calculus problem, someone in primary school could solve, wrong. For users this makes no sense.

[–] november@piefed.blahaj.zone 15 points 3 days ago (1 children)

Amazing. How did technology get so dumb?

[–] MalReynolds@slrpnk.net 13 points 3 days ago

Just the next, yet to be made illegal, Ponzi scheme fueled by greed going off the rails.

[–] MalReynolds@slrpnk.net 5 points 3 days ago* (last edited 3 days ago)

Quite so, thanks for the hearty follow up. I was just demystifying MCP. FWIW Qwen3.5 is downright keen to use tools in my experience.

[–] RustyNova@lemmy.world 7 points 3 days ago

AI for thee but not for me

[–] bitjunkie@lemmy.world 26 points 3 days ago (2 children)

can I make them do extra work that would make their contributions genuinely valuable?

The worst-case scenario for this is that you make them do extra work that isn't genuinely valuable, and it wastes their lazy devs' tokens.

[–] SpaceNoodle@lemmy.world 20 points 3 days ago

You misspelled "best-case"

[–] PattyMcB@lemmy.world 10 points 2 days ago

Hit 'em where it hurts. In the wallet

[–] Wiz@midwest.social 6 points 2 days ago

Unless we figure out how to evolve our processes – which includes being able to recognize and distinguish bot contributions – open-source maintenance is going to grind to a halt.

I wish there was some sort of "trust network" where actual humans can say, "Yes, I know this contributor IRL, and vouch for them." But that's also a security problem. So, 🤷🏼

[–] BradleyUffner@lemmy.world 6 points 2 days ago

This might be the most ironic thing I've ever read...