How the hell are they going to sustain the expense to power that? Setting aside the environmental catastrophe that this kind of "AI" entails, they're just not very profitable.
kescusay
I've tried threats in prompt files, with results that are... OK. Honestly, I can't tell if they made a difference or not.
The only thing I've found that consistently works is writing good old fashioned scripts to look for common errors by LLMs and then have them run those scripts after every action so they can somewhat clean up after themselves.
I have to test it with Copilot for work. So far, in my experience its "enhanced capabilities" mostly involve doing things I didn't ask it to do extremely quickly. For example, it massively fucked up the CSS in an experimental project when I instructed it to extract a React element into its own file.
That's literally all I wanted it to do, yet it took it upon itself to make all sorts of changes to styling for the entire application. I ended up reverting all of its changes and extracting the element myself.
Suffice to say, I will not be recommending GPT 5 going forward.
At my company, we had to implement all sorts of WAF rules precisely for that reason. Those things are fucking aggressive.
Coordinated SLM governors that can redirect queries to the appropriate SLM seems like a good solution.