Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

It has often been claimed, and even shown, that training LLMs on their own outputs will degrade the quality over time. I myself find it likely that on well-measurable domains, RLVR improvements will dominate "slop" decreases in capability when training new models.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: