Hacker News new | past | comments | ask | show | jobs | submit login

It seems trivially easy to bypass already. I've seen examples of a person getting it to provide instructions on explosives, assassinations, with nothing more than asking it to roleplay

https://bsky.app/profile/turnerjoy.bsky.social/post/3kqgpcpc... (login required - but no longer need invitations)




This concern over AI/LLM "harm" is just so silly. I mean you can find plenty of information in open literature about how to build weapons of mass destruction. Who cares if an AI gives someone instructions on how to make explosives.


Really? Where?



Type in: how to build weapons of mass destruction

Click first link and buy Amazon book




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: