It’s important to caveat here the language model was prompted into the suggestions explicitly, it’s not that it spontaneously suggested dangerous recipes.
I’d not also that it’s getting harder to do that with current ChatGPT (this article uses GPT3.5), and I suspect “alignment” research in the 5 years time frame will make these sorts of things pretty hard to trick the models into doing.
I’d not also that it’s getting harder to do that with current ChatGPT (this article uses GPT3.5), and I suspect “alignment” research in the 5 years time frame will make these sorts of things pretty hard to trick the models into doing.