Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

I firmly disagree because this isn't a learning model - it uses the same model for all output. So, unlike Microsoft's Twitter experiments, it can't be abused at scale into spitting out hate. So, any abuse is forced to find flaws in a fixed model and won't affect anyone else using it, which is much harder and pointless. What's the point? It's all for giggles, no actual harm caused, humans can write whatever they want anyway.


This model is not learning, but the transcripts being farmed at scale are for sure being used (with human assistance and control) for building the next models. I don't think the kind of exploit of the microsoft bot is possible, however the next model might give a superficial appearance of being safer, since the transcripts we are given OpenAI of us trying to outsmart the model will be used to train it further.




Consider applying for YC's Winter 2026 batch! Applications are open till Nov 10

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: