468
Jailbroken AI Chatbots Can Jailbreak Other Chatbots
(www.scientificamerican.com)
This is a most excellent place for technology news and articles.
What possible legitimate reason could someone need to know how to make chlorine/mustard gas?
Apart from the fact that they are made from common household products, are easy to make by mistake, and can kill you.
Wait that’s true of napalm as well… fuck.