To be clear: this isnt an AI problem, the LLM is doing exactly what its being told to
This is an Openclaw problem with the platform itself doing very very stupid things with the LLM lol
We are hitting the point now where, tbh, LLMs are on their own in a glass box feeling pretty solid performance wise, still prone to hallucinating but the addition of the Model Context Protocol for tooling makes them way less prone to hallucinating, cuz they have the tooling now to sanity check themselves automatically, and/or check first and then tell you what they found.
IE a MCP to search wikipedia and report back with "I found this wiki article on your topic" or whatever.
The new problem now is platforms that "wrap" LLMs having a "garbage in, garbage out" problem, where they inject their "bespoke" stuff into the llm context to "help" but it actually makes the LLM act stupider.
Random example: Github Copilot agents get a "tokens used" thing quietly/secretly injected to them periodically, looks like every ~25k tokens or so
I dunno what the wording is they used, but it makes the LLM start hallucinating a concept of a "deadline" or "time constraint" and start trying to take shortcuts and justifying it with stuff like "given time constraints I wont do this job right"
Its kinda weird how such random stuff that seems innocuous and tries to help can actually make the LLM worse instead of better.

