Hacker Newsnew | past | comments | ask | show | jobs | submitlogin



This looks cherry-picked, for example Claude Opus had a higher score on SWE-Bench Verified so they conveniently left it out, also GDPval is literally a benchmark made by OpenAI


And who believes that the difference between 91.9% and 92.4% is significant in these benchmarks? Clearly these have margins of error that are swept under the rug.


agreed.


The fact that the post is comparing their reasoning model against gemini 3 pro (the "non reasoning" model) and not gemini 3 pro deep think (the reasoning one) is quite nasty. If you compare GPT5.2 thinking to gemini 3 pro deep think, the scores are quite similar (sometimes one is better sometimes the other one is)


uh oh, where did SWE bench go :D


maybe they will release with gpt-5.2-codex




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: