this post was submitted on 26 Nov 2025
444 points (98.7% liked)
Fuck AI
4701 readers
1404 users here now
"We did it, Patrick! We made a technological breakthrough!"
A place for all those who loathe AI to discuss things, post articles, and ridicule the AI hype. Proud supporter of working people. And proud booer of SXSW 2024.
AI, in this case, refers to LLMs, GPT technology, and anything listed as "AI" meant to increase market valuations.
founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
Genuinely, this is the driving misconception people have about AIs right now. That somehow everybody using them is making them smarter, when really it’s leading to model collapse
Can you help correct this for me? Don't you feed them valuable training data and exposure to real world problems in the process of using them?
No. AI models are pre-trained, they do not learn on the fly. They are hoping to discover General Artificial Intelligence, which is what you are describing. The problem is that they don't even understand exactly how training even works. While engineers understand the overall architecture, the specific "reasoning" or decision-making pathways within the model are too complex to fully interpret, leading to a gap between how it works and why it makes a particular decision.
My assumption wasn't that they learned on the fly, it was that they were trained on previous interactions. Eg the team developing them would use data collected from interaction with model v3 to train model v4. Seems like juicy relevant data they wouldn't even have to go steal and sort
That's true to an extend, but the interactions are only useful for training if you can mark it as good / bad etc (which is why sometimes apps will ask you if they were useful). But the 'best' training data like professional programming etc is usually sold at a premium tier with a promise not to use your data for training (since corporations don't want their secrets getting out).
You can’t train ai on ai output. It causes degradation on the newly trained model.
First: that's wrong, every big LLM uses some data cleaned/synthesized by previous LLMs. You can't solely train on such data without degradation, but that's not the claim.
Second: AI providers very explicitly use user data for training, both prompts and response feedback. There's a reason businesses pay extra to NOT have their data used for training.
yep ai training on ai is totally making things better…
I mean - yeah, it is? This is a well-researched part of the data pipelines for any big model. Some companies even got into trouble because their models identified as other models, whose outputs they were trained on.
It seems you have a specific bone to pick that you attribute to such training, but it's just such a weird approach to deny pretty broadly understood results...
i’m aware i also know it increases hallucinations
No, it doesn't. Unless you can show me a paper detailing that literally any amount of synthetic data increases hallucinations, I'll assume you simply don't understand what you're talking about.
what paper? no one in industry is gonna give you this shit, it's literal gold
academics are still arguing about it but save this and we can revisit in 6 months for a fat i told you so if you still care
ai is dead as shit for anything that matters until this issue is fixed
but at least we can enjoy soulless art while we wait for the acceleration
Yeah, that's what I guessed. Try to look into the research first before making such grandiose claims.
i know the current research, i know it’s going to eat your lunch
Ah yes, and you can't show us that research because it goes to another school? And all companies that train LLMs are simply too stupid to realize this fact? Their research showing the opposite (which has been replicated dozens of times over) was just a fluke?
no because this is literally in development, this isn’t some 60 year old mature tech
algorithms sure, nn for some narrow topics yep great, not the this bullshit though
there is already academic accessible research talking about LLM issues of which the major concern is hallucinations, to the point where the word bailout is starting to make the rounds in the us from these very companies
the argument is whether or not you believe this is inherent or fixable and a big focus is on the training
anyone listening to any ai company right now is a damn fool with the obvious circular vendor bullshit going on
but you do you, if the market could be trusted to be sane i’d be timing it right now
Of course, you don't have research supporting your position because it's still in development. So obviously we can just ignore all the papers released over the last decade+ which show the opposite of what you're claiming - convenient!
Yeah, as I expected - you literally don't understand what this conversation is even about. Since you have a bone to pick with the industry, you make up random claims that you think make the industry look bad. But what you don't understand is: you're just making a fool of yourself by making subjective claims around topics you simply don't understand. Critique the AI industry for the greedy, useless shit they're doing and creating, not by making up wrong "facts" and ignoring all evidence against them.
And just to save us both time, I'll list try to list positions you seem to think I hold, which I don't:
If you choose to reply again and think I'm lying about not holding these positions, re-read the conversation until you understand it.
the focus i know about is literally on this issue, you keep acting like you’d actually expect anyone to give you details lol
but sure please give me your facts and access to your research