Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

> The easiest way to inoculate yourself against a persuasive technique is to be aware of it ahead of time.

Or you can avoid being exposed to it. If you think you know all the techniques an AI might use against you, you're less likely to do that.

The point of the experiment isn't "let's work out how an AI might try to persuade us to let it out". It's "even a human intelligence can persuade people who think they could never be persuaded, do you really trust yourself to do better against a superhuman one?"

If you don't know why the gatekeeper failed, it's harder to come up with bullshit reasons why you would have succeeded in that position.



Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: