Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

What do you know, the human results line up exactly with ChatGPT. What are the odds! Surely the human responders are highly ethical individuals and they wouldn't even dream of copy-pasting all the questions into ChatGPT without reading them.

Realistically, this mostly tells me that the "human answers" service is dead. People will figure out a way to pass the work off to an AI, regardless of quality, as long as they can still get paid.



Yea funny coincidence, but this is not at all how the human answers were collected.

Rapidata answered this in another comment below. They integrate micro-surveys into mobile apps (like Duolingo, games, etc) as an optional opt-in instead of watching ads. The users are vetted and there's no incentive to answer correctly.


Yeah, I always intentionally choose a wrong answer when I get one of those kinds of ads. Little acts of rebellion.


But, there is a clear incentive to answer the question incorrectly. The wrong answer is funny and will give the human some level of pleasure thinking about it. I would certainly reply with "walk" just for fun and apparently 28.5% of people agree with me.


In which case the %age is notable as it aligns very closely to the effect size on cookie accept/reject ratios. Cookie dialogs tend to fall 70/30 either way.


> there's no incentive to answer correctly

Answering correctly is not in question here. This is essentially opinion polling anyway, there is no single correct answer.

The incentive is exactly what you said: to skip ads.

How are the users actually vetted? We have no information on this, just have to take rapidata on faith.


> there is no single correct answer

I think we all mostly agree that there is a single correct answer, and that is why this discussion exists in the first place.


The default model of ChatGPT is GPT 5.2 Instant, not the one lines up with human results (which is GPT 5).

However, it does tell us something about human answers as the above commenter confidently reached such a strong but baseless conclusion.


thats almost always been the case with 3rd party human task services


Yup. I was surprised that the article author took the results at face value. Having results that match the most commonly-known AI platform's results perfectly seemed worthy of a mention!




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: