Hacker News new | past | comments | ask | show | jobs | submit login

Neural LMs used to be based on recurrent architectures until the Transformer came along. That architecture is not recursive.

I am not sure that a diffusion approach is all that suitable for generating language. Word are much more discrete than pixels.






I meant sequential generation, I didn't mean using an RNN.

Diffusion doesn't work on pixels directly either, it works on a latent representation.


All NNs work on latent representations.

The contrast here is real: there are pixel space diffusion models and latent space diffusion models. Pixel space diffusion is slower because there's more redundant information.



Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: