The researchers are using a way named adversarial training to stop ChatGPT from allowing people trick it into behaving badly (often known as jailbreaking). This function pits a number of chatbots from one another: a single chatbot plays the adversary and attacks An additional chatbot by generating textual content to drive it to buck its normal cons