Hacker Newsnew | past | comments | ask | show | jobs | submitlogin
Just Ask for Generalization (2021) (evjang.com)
38 points by jxmorris12 13 days ago | hide | past | favorite | 4 comments





(2021), still very interesting. Especially the "post-overfitting" training strategy is unexpected.

This is talking about the double descent phenomenon (https://en.wikipedia.org/wiki/Double_descent)

I remember vaguely that this was observed when training GPT-3 (probably?) as well. Just trained on and on, and the error went up and then down again. Like a phase transition in the model.

The low sample efficiency of RL is well explained.



Consider applying for YC's Fall 2025 batch! Applications are open till Aug 4

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: