this post was submitted on 30 Nov 2025
41 points (97.7% liked)
Fuck AI
4728 readers
535 users here now
"We did it, Patrick! We made a technological breakthrough!"
A place for all those who loathe AI to discuss things, post articles, and ridicule the AI hype. Proud supporter of working people. And proud booer of SXSW 2024.
AI, in this case, refers to LLMs, GPT technology, and anything listed as "AI" meant to increase market valuations.
founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
thank you for your response, the cauliflower anecdote was enlightening. your description of it being a statistical prediction model is essentially my existing conception of LLMs, but this was only really from gleaning other's conceptions online, and I've recently been concerned it was maybe an incomplete simplification of the process. I will definitely read up on markov chains to try and solidify my understanding of LLM 'prediction
I have kind of a follow up if you have the time. I hear a lot that LLMs are "running out of data" to train on. When it comes creating a bicycle schematic, it doesn't seem like additional data would make an LLM more effective at a task like this, since its already producing a broken amalgamation. It seems like generally these shortcomings of LLMs' generalizations would not be alleviated by increased training data. So what exactly is being optimized by massive increases (at this point) in training data--or, conversely, what is threatened by a limited pot?
I ask this because lots of people who preach that LLMs are doomed/useless seem to focus in on this idea that their training is limited. To me their generalization seems like evidence enough that we are no where near the tech-bro dreams of AGI.
No, you're quite correct: Additional training data might increase the potential for novel responses and thus enhance the perception of apparent creativity, but that's just another way to say "decrease correctness". To stick with the example, if you wanted to have an LLM yield a better bicycle, you should if anything be partitioning the training data and curating it. Garbage in, garbage out. Mess in, mess out.
Put it another way: Novelty implies surprise, surprise implies randomness. Correctness implies consistently yielding the solitary correct answer. The two are inherently mutually opposed.
If you're interested in how all this nonsense got started, I highly recommend going back and reading Weizenbaum's original 1966 paper on ELIZA. Even back then, he knew better:
Weizenbaum quickly discovered the harmful effects of human interactions with these kinds of models:
god, the reactions to eliza is such a harbinger of doom. real cassandra moment. it's an extra weird touchstone for me because we had it on our school computers in the late 90s. the program was called
DOCTORand basically behaved identically to the original, eg find a noun and use it in a sentence. as a 9-year old i found it to be ass, but i've only recently learned that some people anthropomorphise everything and can lose themselves totally in "tell me about boats" even if they rationally know what the program is actually doing.as a 30-something with some understanding of natural language processing, eliza is quite nifty.