misk@sopuli.xyz to Technology@lemmy.worldEnglish · 1 year agoJailbroken AI Chatbots Can Jailbreak Other Chatbotswww.scientificamerican.comexternal-linkmessage-square70fedilinkarrow-up1386arrow-down114
arrow-up1372arrow-down1external-linkJailbroken AI Chatbots Can Jailbreak Other Chatbotswww.scientificamerican.commisk@sopuli.xyz to Technology@lemmy.worldEnglish · 1 year agomessage-square70fedilink
minus-squareKairuByte@lemmy.dbzer0.comlinkfedilinkEnglisharrow-up4·1 year agoWhat possible legitimate reason could someone need to know how to make chlorine/mustard gas? Apart from the fact that they are made from common household products, are easy to make by mistake, and can kill you. Wait that’s true of napalm as well… fuck.
What possible legitimate reason could someone need to know how to make chlorine/mustard gas?
Apart from the fact that they are made from common household products, are easy to make by mistake, and can kill you.
Wait that’s true of napalm as well… fuck.