Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Amazing! OpenAI figured out how to scale inference. https://arxiv.org/abs/2407.21787 show how using more compute during inference can outperform much larger models in tasks like math problems

I wonder how do they decide when to stop these Chain of Thought for each query? As anyone that played with agents can attest, LLMs can talk with themselves forever.



Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: