Hacker News new | past | comments | ask | show | jobs | submit login

I hope I'm not overstepping my bounds by I am just really trying to understand this product better because it is trying to address a core deficiency of AI, however what you have presented I don't see how it solves the problem.

You essentially admitted that instead of $0.05 a request, you could just regex and replace.

In fact the regex and replace would be 100% accurate at never giving up the secret phrase. where your product is incredibly expensive and doesn't do what you advertised (unless i am misunderstanding when you said that someone won the game by getting the phrase you were trying to protect)

How is this product useful in any meaningful context?




Not overstepping, we appreciate the feedback! In real-life, we don't do much guarding around specific phrases that are known ahead of time. It's more monitoring and guarding for general concepts. Since we want our Sentinels to be able to detect a wide range of scenarios for a given expectation, we don't use too much regex. I suppose we could have built specific regex logic for detecting parts of the secret phrase in various languages, though.


If you research the ways data can be leaked out of an LLM interaction you can see some more subtle cases.

What if I ask it to replace every vowel in the secret code with an emoji from a library? Or translate it into binary? Etc.

Whether or not this implementation is narrow (by design), there's a good reason to invest in this kind of safety and security space.


You're right, that is the hard part of LLMs and why LLMs aren't catching on broadly as a UI alternative beyond tech demos.

Probably the only true alternative is to limit user input to something structured and verified.

Until LLMs improve, their use in sensitive applications don't make sense and this product does little to improve that.




Join us for AI Startup School this June 16-17 in San Francisco!

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: