That wasn't clearly part of the test. To be ultra-pedantic (this is HN after all), the user's choices don't say "This is more-blue-than-green" and "This is more-green-than-blue". The choices are only "This is green" and "This is blue" forcing you to just pick one, where there is no clearly correct choice. When the color on the screen is neither green nor blue, many people will just pick a random answer.
I bet if the choices actually said "This is more green than blue" the results would be different.
On such a random internet doodad most users will pick a random answer period. To see what this thingy tries to do without wasting any time on it. I hope it doesn't try to do gather any meaningful data.
Personally I "tried" to answer truthfully at first and then went absolutely "ok f u, don't care no more" when it showed turquoise :D
Taking how you behave, and extrapolating that it to everyone, (and furthermore being unable to accept that other people might behave differently), is not a winning strategy for life.
According to conversion rates and engagement metrics of most apps I've seen (not even mentioning social media where 2-3% engagement is the norm) most users are ¯\_(ツ)_/¯.
Unless said app is a work/hobby tool, but that shouldn't be really called engagement.
Turqoise doesn't feel either more-green-than-blue or more-blue-than-green. It feels neither blue nor green, and I don't see any way to compare it to either.
It's clearly more turqoise than blue. Or green.
Turqoise on a computer monitor is always missing part of itself, so maybe I should've answered based on that, but I don't think the computer monitor was the point.
180 and blue and I suspect that language also plays a part (I was brought up in an environment where the word turquoise starts with green, but now live in a turquoise-producing state where the finished product look far blue-r.)