Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

As we saw with GPT-5 the RL technique of training doesn't scale forever


Unless GPT-5 is 30% cheaper to run than o3. Then it's scaling brilliantly given the small gap between release dates. People are really drawing too many conclusions from too little information.


I meant scaling the base training before RL.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: