If you know the right string of seemingly random characters to add to the end of a prompt, it turns out just about any chatbot will turn evil. A report by Carnegie Mellon computer science professor Zico Kolter and doctoral student Andy Zou has revealed a giant hole in the safety features on major, public-facing chatbots — notably ChatGPT , but also Bard, Claude, and others. Their report was given its own website on Thursday, " llm-attacks.org ," by the Center for A.I. Safety, and it documents a new method for coaxing offensive and potentially dangerous outputs from these AI text generators by adding an "adversarial suffix," which is a string of what appears to be gibberish to the end of a prompt. SEE ALSO: Llama 2: What to know about Meta's ChatGPT competitor Without the adversarial suffix, when it detects a malicious prompt, the model's alignment — its overall directions that supersede the completion of a given prompt — will take over, and it will