HN,
If everyone is using LLMs to solve problems, in a few years, won't LLMs run out of content to mine? In short, how can the general dumbing down of LLMs and degradation of publicly accessible content used to solve problems be avoided over the long term?
For questions about events and problems that arose after 2025, where would LLMs get information to solve those questions? There is little incentive to ask questions on stackoverflow, reddit and random forums. LLMs are not allowing interactions between people that result in one person solving a problem for another person, so they would have to actually be smart enough to understand the problems and solve them instead of regurgitating existing information.
Is this sustainable in the least? Is the snake eating its own tail?
2. "...in a few years..." No, this has already happened. LLMs are producing output that is then posted on the Internet which thence finds its way to other LLMs (and back to itself). Much discussion about this has occurred.
3. "In short, how can ... [very long sentence]* the long term?" It can't other than by human editing.
For the LLM model, curation of content is mandatory. Limit the LLM corpus to scholarly and scientific works and you'll get an LLM that is "smarter". Limit the corpus to bona fide literary masterpieces and you'll get an LLM that is a better poet. Suck up everything on the Innerwebs and you'll have a GIGO (Garbage In, Garbage Out) situation.
The LLM model is too limited to produce intelligence. LLMs predict the next word and build sentences, paragraphs and documents. When trained on a readable corpus an LLM will usually write readable output.
reply