Hacker News new | past | comments | ask | show | jobs | submit login

I think it's reasonable to say that this was actually the point of releasing LLMs publicly. The companies that created them wanted a moat and figured the data they had could be it if they poisoned anyones attempt to collect the same data in the future.



Intriguing thought, but arguably people are intentionally using GPT to generate synthetic data for their domain specific model. So I'm kinda torn between AI giants poisoning the well with their models, or it just being unforeseen consequences (or one they willingly ignored to be first to market).


Generating data from ChatGPT even $1,000,000 in tokens worth can’t be on the same scale as what OpenAI is collecting from everyone.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: