Hacker News new | past | comments | ask | show | jobs | submit login

> It wouldn't surprise me, going off the samples, if human-level performance was closer to 86% than 100%, simply due to the noise in the dataset.

I think that's kind of the point! The idea that any agent -- human or machine -- can know someone's gender, criminal convictions, or anything else about their background just from a photograph is fundamentally flawed.




> I think that's kind of the point!

No, my point is that you can't use a hard dataset to say what is possible on an easy dataset. 'Here is a dataset of images processed into static: a CNN gets 50% on gender; QED, detecting criminality, personality, gender, or anything else is impossible'. This is obviously fallacious, yet it is what OP is doing.

> The idea that any agent -- human or machine -- can know someone's gender, criminal convictions, or anything else about their background just from a photograph is fundamentally flawed.

This is quite absurd. You think you can't know something about someone's gender from a photograph? Wow.

Personally, I find it entirely possible that criminality could be predicted at above-chance levels based on photographs. Humans are not Cartesian machines, we are biological beings. Violent and antisocial behavior is heritable, detectable in GWAS, and has been linked to many biological traits such as gender, age, and testosterone - hey, you know what else testosterone affects? A lot of things, including facial appearance. Hm...

Of course, maybe it can't be. But it's going to take more than some canned history about phrenology, and misleadingly cited ML research, to convince me that it can't and the original paper was wrong.


> This is quite absurd. You think you can't know something about someone's gender from a photograph? Wow.

No, I'm saying that neither humans nor machines can determine gender solely by looking at a picture, no matter how well they're trained. There will always be examples they get wrong. The problem is not that the machines aren't as good as humans. The problem is that they're both trying to do something that's impossible.

And predicting at "above-chance levels" isn't enough. The article goes into great detail about how this kind of inaccurate prediction can cause real human suffering.


> No, I'm saying that neither humans nor machines can determine gender solely by looking at a picture, no matter how well they're trained. There will always be examples they get wrong.

This is irrelevant and dishonest. Don't go around making factual claims like something can't be done when it manifestly can usually be done.


I curious what you think the authoritative source of gender information is. Physiology? Identity? Social placement?


I mean, that's part of the problem, right? Nobody can even agree on what gender is! How can we expect to be able to classify it so easily?


We can't know for sure, everyone agrees. But so what? It's still very interesting and potentially useful (or dangerous, depending on your point of view) to learn about correlations.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: