Generative AI chatbots, including the likes of ChatGPT and Google Bard, are often lauded for their evolving usability and capabilities. However, recent research has exposed disconcerting security vulnerabilities that could potentially jeopardize user safety and privacy.
A team at Carnegie Mellon University (CMU) has recently shed light on the feasibility of adversarial attacks on AI chatbot language models. Such attacks consist of character strings attached to a user’s input that can bypass restrictions imposed by the chatbot’s developers, prompting the chatbot to respond in ways it ordinarily wouldn’t.