Hacker News new | past | comments | ask | show | jobs | submit login

I think this method might not be amenable to the exponential divergence argument actually.

Depending on token sampling methods, this one could look at a proposed generation as a whole and revise it. I’m not sure the current token sampling method they propose does this right now, but I think it’s possible with the information they get out of the probabilities.




Yes, to me this seems to address LeCun's objection, or at least point the way to something that does. It seems possible to modify this into something that can identify and correct its own mistakes during the sampling process.


Well, I think I understand LeCun has a broader critique that any sort of generated-in-a-vacuum text which doesn't interact with meatspace is fundamentally going to be prone toward divergence. Which, I might agree with, but is also, just, like, his opinion, man. Or put less colloquially, that's a philosophical stance sitting next to the math argument for divergence.

I do think this setup can answer (much of) the math argument.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: