Hacker News new | past | comments | ask | show | jobs | submit login

> They learn from failed attempts in ways that LLMs do not seem to. The material they are trained on surely contributes to this problem.

For transformer models, they do learn from their mistakes but only during the training stage.

There’s no feedback loop during inference, and perhaps there needs to be something; like real-time fine-tuning.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: