From the repo:
Have opinions. Don't just report facts - react to them. "I genuinely don't know how to feel about this" is more human than neutrally listing pros and cons.
"We did it, Patrick! We made a technological breakthrough!"
A place for all those who loathe AI to discuss things, post articles, and ridicule the AI hype. Proud supporter of working people. And proud booer of SXSW 2024.
AI, in this case, refers to LLMs, GPT technology, and anything listed as "AI" meant to increase market valuations.
From the repo:
Have opinions. Don't just report facts - react to them. "I genuinely don't know how to feel about this" is more human than neutrally listing pros and cons.
Stuff like that doesn't always work though, at least on free versions in my experience. I use Ai to write flowery emails to people to sound nice when I normally wouldn't bother and I used it to negotiate buying my car. I would continually tell it not to use - dashes while writing emails. And inevitably after 1 answer it would go back to using them.
Maybe paid versions are different but on free ones you have to continually correct it.
I really despise how Claude's creators and users are turning the definition of "skill" from "the ability to use [learned] knowledge to enhance execution" into "a blurb of text that [usefully] constrains a next-token-predictor".
I guess, if you squint, it's akin to how biologists will talk about species "evolving to fit a niche" amongst themselves or how physicists will talk about nature "abhorring a vacuum". At least they aren't talking about a fucking product that benefits from hype to get sold.
I can't help but get secondhand embarrassment whenever I see someone unironically call themselves a "prompt engineer". 🤮
Hey, they had to learn thermodynamics and spend 3 semesters in calculus to write those prompts
Isn't this a thing that authoritarians do. They co-opt language. It's the same thing conservatives do. The venn diagram of tech bros and the far right is too close to being a circle.
You can pretty put any word out of the dictionary into a search engine and the first results are some tech company that took the word either as their company name or redefined it into some buzzword.
Skills were functions/frameworks built for Alexa, so they just appropriated the term from there.
You do understand this is more akin to white hat testing, right?
Those who want to exploit this will do it anyway, except they won't publish the result. By making the exploit public, the risk will be known if not mitigated.
I'm admittedly not knowledgeable in White Hat Hacking, but are you supposed to publicize the vulnerability, release a shortcut to exploit it telling people to 'enjoy', or even call the vulnerability handy ?
Responsible disclosure is what a white hat does. You report the bug to whomever is the party responsible for patching and give them time to fix it.
That sort of depends on the situation. Responsible disclosure is for if there is some relevant security hole that is an actual risk to businesses and people, while this here is just "haha look LLMs can now better pretend to write good text if you tell it to". That's not really responsible disclosurable. It's not even specific to one singular product.
Considering the "vulnerability" here is on the level of "don't use password as your password" - yeah, releasing it all is exactly the right step.
If these "signs of AI writing" are merely linguistic, good for them. This is as accurate as a lie detector (i.e., not accurate) and nobody should use this for any real world decision-making.
The real signs of AI writing are not as easy to fix as just instructing an LLM to "read" an article to avoid them.
As a teacher, all of my grading is now based on in person performances, no tech allowed. Good luck faking that with an LLM. I do not mind if students use an LLM to better prepare for class and exams. But my impression so far is that any other medium (e.g., books, youtube explanation videos) leads to better results.
Bro isnt even gonna check its output anyway.
How likely is the list to be AI generated as well?
Wikipedia is one of the last genuine places on the Internet, and these rat bastards are trying to contaminate that, too
Wikipedia just sold the rights to use Wikipedia for AI training to Microsoft and openai....
It's getting scraped anyway. So why not get some money from it?
They lose the right to sue them
They probably realized that it was a losing battle and they didn't want to pay legal fees.
Imo this. Selling access also implies its illegal to access without purchasing rights which imho helps undermine AI's only monetary advantage
How exactly does that work? Wikipedia does not "own" the content on the website, it's all CC-BY licensed.
Yeah, they're selling the work of others. That's how the site always worked. This venture into "AI" is nothing new.
The BY term is not respected by LLMs
So? Still doesn't make sense to me that wikipedia can sell anything meaningful here, but I'm also not a lawyer. Do they promise not to sue them or sell them some guarantee that contributors also can't sue them? Is it just some symbolic PR washing?
Seeing as OpenAI struggled to make its AI avoid the em dash and still hasn't entirely managed to do it, I'm not too worried.
You have to understand that their public facing product is not the same as the one they allow enterprise or state actors to use.
They benefit from public thinking they have these stupid limitations, gives them more space to curate their product offerings where the real money is made.
I don't understand how the public thinking these are bad products is an incentive for especially state actors to use them. That seems counterintuitive.
TBF OpenAI are a bunch of idiots running the world's largest ponzi scheme. If DeepMind tried it and failed then...
Well I still wouldn't be surprised, but at least it would be worth citing.
It can't avoid doing those things. That's the reason for the article.
Wikipedia is astroturfed BS for anything remotely politically related.
Useful if you want to learn about the Ivory-billed Woodpecker or a closed-cycle regenerative heat engine, etc..
So no politics and subjects with political implications such as history.
It's really "strange" how people have this delusional viewpoint of wikipedia as neutral, honest, etc.
I guess it seems fine if you've got lib politics completely within the hegemonic narrative.
That's what prolewiki is for
I trust no one with my history or geopolitics but ML's.
Most informed, well read and thorough people on the planet